Fold : 0
[I 2020-09-27 04:38:19,970] A new study created in memory with name: no-name-4ce5d708-0a14-4cfc-bc23-b74b1ae8bc8d
feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000784 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.574988 valid's binary_logloss: 0.650038
[200] train's binary_logloss: 0.51996 valid's binary_logloss: 0.651275
Early stopping, best iteration is:
[137] train's binary_logloss: 0.552891 valid's binary_logloss: 0.649537
feature_fraction, val_score: 0.649537: 14%|#4 | 1/7 [00:00<00:04, 1.41it/s][I 2020-09-27 04:38:20,696] Trial 0 finished with value: 0.6495366523489564 and parameters: {'feature_fraction': 0.7}. Best is trial 0 with value: 0.6495366523489564.
feature_fraction, val_score: 0.649537: 14%|#4 | 1/7 [00:00<00:04, 1.41it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000262 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.579394 valid's binary_logloss: 0.650899
Early stopping, best iteration is:
[92] train's binary_logloss: 0.584248 valid's binary_logloss: 0.650087
feature_fraction, val_score: 0.649537: 29%|##8 | 2/7 [00:01<00:03, 1.58it/s][I 2020-09-27 04:38:21,152] Trial 1 finished with value: 0.6500867396462259 and parameters: {'feature_fraction': 0.5}. Best is trial 0 with value: 0.6495366523489564.
feature_fraction, val_score: 0.649537: 29%|##8 | 2/7 [00:01<00:03, 1.58it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000583 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.583335 valid's binary_logloss: 0.652707
Early stopping, best iteration is:
[88] train's binary_logloss: 0.590815 valid's binary_logloss: 0.652179
feature_fraction, val_score: 0.649537: 43%|####2 | 3/7 [00:01<00:02, 1.77it/s][I 2020-09-27 04:38:21,556] Trial 2 finished with value: 0.652179314698687 and parameters: {'feature_fraction': 0.4}. Best is trial 0 with value: 0.6495366523489564.
feature_fraction, val_score: 0.649537: 43%|####2 | 3/7 [00:01<00:02, 1.77it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000873 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.573386 valid's binary_logloss: 0.65332
Early stopping, best iteration is:
[72] train's binary_logloss: 0.592163 valid's binary_logloss: 0.652672
feature_fraction, val_score: 0.649537: 57%|#####7 | 4/7 [00:02<00:01, 1.90it/s][I 2020-09-27 04:38:21,994] Trial 3 finished with value: 0.6526718623826693 and parameters: {'feature_fraction': 0.8}. Best is trial 0 with value: 0.6495366523489564.
feature_fraction, val_score: 0.649537: 57%|#####7 | 4/7 [00:02<00:01, 1.90it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000825 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.573309 valid's binary_logloss: 0.653575
Early stopping, best iteration is:
[67] train's binary_logloss: 0.59526 valid's binary_logloss: 0.652553
feature_fraction, val_score: 0.649537: 71%|#######1 | 5/7 [00:02<00:01, 1.58it/s][I 2020-09-27 04:38:22,880] Trial 4 finished with value: 0.6525525222524156 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 0 with value: 0.6495366523489564.
feature_fraction, val_score: 0.649537: 71%|#######1 | 5/7 [00:02<00:01, 1.58it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.006985 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.576836 valid's binary_logloss: 0.653247
Early stopping, best iteration is:
[70] train's binary_logloss: 0.596318 valid's binary_logloss: 0.651989
feature_fraction, val_score: 0.649537: 86%|########5 | 6/7 [00:03<00:00, 1.79it/s][I 2020-09-27 04:38:23,260] Trial 5 finished with value: 0.6519889736642206 and parameters: {'feature_fraction': 0.6}. Best is trial 0 with value: 0.6495366523489564.
feature_fraction, val_score: 0.649537: 86%|########5 | 6/7 [00:03<00:00, 1.79it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000890 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.571148 valid's binary_logloss: 0.654809
Early stopping, best iteration is:
[46] train's binary_logloss: 0.611505 valid's binary_logloss: 0.653731
feature_fraction, val_score: 0.649537: 100%|##########| 7/7 [00:03<00:00, 1.96it/s][I 2020-09-27 04:38:23,655] Trial 6 finished with value: 0.6537305766184812 and parameters: {'feature_fraction': 1.0}. Best is trial 0 with value: 0.6495366523489564.
feature_fraction, val_score: 0.649537: 100%|##########| 7/7 [00:03<00:00, 1.91it/s]
num_leaves, val_score: 0.649537: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000620 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.57822 valid's binary_logloss: 0.649469
Early stopping, best iteration is:
[85] train's binary_logloss: 0.587383 valid's binary_logloss: 0.649227
num_leaves, val_score: 0.649227: 5%|5 | 1/20 [00:00<00:09, 2.10it/s][I 2020-09-27 04:38:24,142] Trial 7 finished with value: 0.6492270930479551 and parameters: {'num_leaves': 30}. Best is trial 7 with value: 0.6492270930479551.
num_leaves, val_score: 0.649227: 5%|5 | 1/20 [00:00<00:09, 2.10it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000836 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.514021 valid's binary_logloss: 0.655753
Early stopping, best iteration is:
[46] train's binary_logloss: 0.579148 valid's binary_logloss: 0.653741
num_leaves, val_score: 0.649227: 10%|# | 2/20 [00:01<00:08, 2.04it/s][I 2020-09-27 04:38:24,666] Trial 8 finished with value: 0.6537410492395737 and parameters: {'num_leaves': 60}. Best is trial 7 with value: 0.6492270930479551.
num_leaves, val_score: 0.649227: 10%|# | 2/20 [00:01<00:08, 2.04it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.003688 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.635569 valid's binary_logloss: 0.650064
[200] train's binary_logloss: 0.617663 valid's binary_logloss: 0.647802
[300] train's binary_logloss: 0.603448 valid's binary_logloss: 0.648074
Early stopping, best iteration is:
[204] train's binary_logloss: 0.617119 valid's binary_logloss: 0.647726
num_leaves, val_score: 0.647726: 15%|#5 | 3/20 [00:01<00:08, 2.08it/s][I 2020-09-27 04:38:25,128] Trial 9 finished with value: 0.6477259914776615 and parameters: {'num_leaves': 8}. Best is trial 9 with value: 0.6477259914776615.
num_leaves, val_score: 0.647726: 15%|#5 | 3/20 [00:01<00:08, 2.08it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.003613 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.295966 valid's binary_logloss: 0.669876
Early stopping, best iteration is:
[28] train's binary_logloss: 0.505319 valid's binary_logloss: 0.658492
num_leaves, val_score: 0.647726: 20%|## | 4/20 [00:02<00:11, 1.33it/s][I 2020-09-27 04:38:26,505] Trial 10 finished with value: 0.6584916550281572 and parameters: {'num_leaves': 229}. Best is trial 9 with value: 0.6477259914776615.
num_leaves, val_score: 0.647726: 20%|## | 4/20 [00:02<00:11, 1.33it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.003890 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.614941 valid's binary_logloss: 0.650113
[200] train's binary_logloss: 0.584546 valid's binary_logloss: 0.649317
[300] train's binary_logloss: 0.558777 valid's binary_logloss: 0.650116
Early stopping, best iteration is:
[244] train's binary_logloss: 0.57318 valid's binary_logloss: 0.649088
num_leaves, val_score: 0.647726: 25%|##5 | 5/20 [00:03<00:11, 1.29it/s][I 2020-09-27 04:38:27,335] Trial 11 finished with value: 0.6490878119710577 and parameters: {'num_leaves': 15}. Best is trial 9 with value: 0.6477259914776615.
num_leaves, val_score: 0.647726: 25%|##5 | 5/20 [00:03<00:11, 1.29it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000790 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.646526 valid's binary_logloss: 0.653926
[200] train's binary_logloss: 0.634088 valid's binary_logloss: 0.649382
[300] train's binary_logloss: 0.62524 valid's binary_logloss: 0.648291
[400] train's binary_logloss: 0.617685 valid's binary_logloss: 0.648421
Early stopping, best iteration is:
[324] train's binary_logloss: 0.623449 valid's binary_logloss: 0.648061
num_leaves, val_score: 0.647726: 30%|### | 6/20 [00:04<00:10, 1.34it/s][I 2020-09-27 04:38:28,014] Trial 12 finished with value: 0.6480614052342452 and parameters: {'num_leaves': 5}. Best is trial 9 with value: 0.6477259914776615.
num_leaves, val_score: 0.647726: 30%|### | 6/20 [00:04<00:10, 1.34it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008792 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.414675 valid's binary_logloss: 0.664247
Early stopping, best iteration is:
[26] train's binary_logloss: 0.571385 valid's binary_logloss: 0.659552
num_leaves, val_score: 0.647726: 35%|###5 | 7/20 [00:05<00:09, 1.37it/s][I 2020-09-27 04:38:28,706] Trial 13 finished with value: 0.6595522413168151 and parameters: {'num_leaves': 122}. Best is trial 9 with value: 0.6477259914776615.
num_leaves, val_score: 0.647726: 35%|###5 | 7/20 [00:05<00:09, 1.37it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000826 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.470281 valid's binary_logloss: 0.657387
Early stopping, best iteration is:
[37] train's binary_logloss: 0.570263 valid's binary_logloss: 0.654921
num_leaves, val_score: 0.647726: 40%|#### | 8/20 [00:05<00:08, 1.36it/s][I 2020-09-27 04:38:29,454] Trial 14 finished with value: 0.654921475007968 and parameters: {'num_leaves': 85}. Best is trial 9 with value: 0.6477259914776615.
num_leaves, val_score: 0.647726: 40%|#### | 8/20 [00:05<00:08, 1.36it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004565 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.635569 valid's binary_logloss: 0.650064
[200] train's binary_logloss: 0.617663 valid's binary_logloss: 0.647802
[300] train's binary_logloss: 0.603448 valid's binary_logloss: 0.648074
Early stopping, best iteration is:
[204] train's binary_logloss: 0.617119 valid's binary_logloss: 0.647726
num_leaves, val_score: 0.647726: 45%|####5 | 9/20 [00:06<00:08, 1.24it/s][I 2020-09-27 04:38:30,428] Trial 15 finished with value: 0.6477259914776615 and parameters: {'num_leaves': 8}. Best is trial 9 with value: 0.6477259914776615.
num_leaves, val_score: 0.647726: 45%|####5 | 9/20 [00:06<00:08, 1.24it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000847 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.291012 valid's binary_logloss: 0.673269
Early stopping, best iteration is:
[36] train's binary_logloss: 0.469537 valid's binary_logloss: 0.661474
num_leaves, val_score: 0.647726: 50%|##### | 10/20 [00:07<00:09, 1.07it/s][I 2020-09-27 04:38:31,655] Trial 16 finished with value: 0.661474106398792 and parameters: {'num_leaves': 233}. Best is trial 9 with value: 0.6477259914776615.
num_leaves, val_score: 0.647726: 50%|##### | 10/20 [00:07<00:09, 1.07it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007106 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.35312 valid's binary_logloss: 0.66936
Early stopping, best iteration is:
[30] train's binary_logloss: 0.529202 valid's binary_logloss: 0.659327
num_leaves, val_score: 0.647726: 55%|#####5 | 11/20 [00:08<00:08, 1.12it/s][I 2020-09-27 04:38:32,457] Trial 17 finished with value: 0.6593265762527931 and parameters: {'num_leaves': 171}. Best is trial 9 with value: 0.6477259914776615.
num_leaves, val_score: 0.647726: 55%|#####5 | 11/20 [00:08<00:08, 1.12it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.003451 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.516771 valid's binary_logloss: 0.654055
Early stopping, best iteration is:
[73] train's binary_logloss: 0.545672 valid's binary_logloss: 0.65329
num_leaves, val_score: 0.647726: 60%|###### | 12/20 [00:09<00:06, 1.25it/s][I 2020-09-27 04:38:33,041] Trial 18 finished with value: 0.6532896945252986 and parameters: {'num_leaves': 59}. Best is trial 9 with value: 0.6477259914776615.
num_leaves, val_score: 0.647726: 60%|###### | 12/20 [00:09<00:06, 1.25it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000421 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.642488 valid's binary_logloss: 0.652362
[200] train's binary_logloss: 0.628687 valid's binary_logloss: 0.649023
[300] train's binary_logloss: 0.617877 valid's binary_logloss: 0.649151
Early stopping, best iteration is:
[279] train's binary_logloss: 0.620188 valid's binary_logloss: 0.648624
num_leaves, val_score: 0.647726: 65%|######5 | 13/20 [00:10<00:05, 1.31it/s][I 2020-09-27 04:38:33,727] Trial 19 finished with value: 0.6486244754568631 and parameters: {'num_leaves': 6}. Best is trial 9 with value: 0.6477259914776615.
num_leaves, val_score: 0.647726: 65%|######5 | 13/20 [00:10<00:05, 1.31it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009248 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.423908 valid's binary_logloss: 0.659234
Early stopping, best iteration is:
[44] train's binary_logloss: 0.527404 valid's binary_logloss: 0.655427
num_leaves, val_score: 0.647726: 70%|####### | 14/20 [00:11<00:05, 1.07it/s][I 2020-09-27 04:38:35,066] Trial 20 finished with value: 0.6554267059223545 and parameters: {'num_leaves': 116}. Best is trial 9 with value: 0.6477259914776615.
num_leaves, val_score: 0.647726: 70%|####### | 14/20 [00:11<00:05, 1.07it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004318 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.669306 valid's binary_logloss: 0.670416
[200] train's binary_logloss: 0.659732 valid's binary_logloss: 0.660321
[300] train's binary_logloss: 0.654518 valid's binary_logloss: 0.655319
[400] train's binary_logloss: 0.651377 valid's binary_logloss: 0.652506
[500] train's binary_logloss: 0.649375 valid's binary_logloss: 0.650671
[600] train's binary_logloss: 0.648029 valid's binary_logloss: 0.649863
[700] train's binary_logloss: 0.647077 valid's binary_logloss: 0.649248
[800] train's binary_logloss: 0.64636 valid's binary_logloss: 0.649015
[900] train's binary_logloss: 0.645776 valid's binary_logloss: 0.649025
[1000] train's binary_logloss: 0.645301 valid's binary_logloss: 0.649111
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.645301 valid's binary_logloss: 0.649111
num_leaves, val_score: 0.647726: 75%|#######5 | 15/20 [00:12<00:05, 1.02s/it][I 2020-09-27 04:38:36,284] Trial 21 finished with value: 0.649110563220017 and parameters: {'num_leaves': 2}. Best is trial 9 with value: 0.6477259914776615.
num_leaves, val_score: 0.647726: 75%|#######5 | 15/20 [00:12<00:05, 1.02s/it][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011497 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.560366 valid's binary_logloss: 0.653399
[200] train's binary_logloss: 0.496525 valid's binary_logloss: 0.653927
Early stopping, best iteration is:
[123] train's binary_logloss: 0.544369 valid's binary_logloss: 0.65141
num_leaves, val_score: 0.647726: 80%|######## | 16/20 [00:13<00:03, 1.10it/s][I 2020-09-27 04:38:36,924] Trial 22 finished with value: 0.6514101116698465 and parameters: {'num_leaves': 38}. Best is trial 9 with value: 0.6477259914776615.
num_leaves, val_score: 0.647726: 80%|######## | 16/20 [00:13<00:03, 1.10it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005527 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.669306 valid's binary_logloss: 0.670416
[200] train's binary_logloss: 0.659732 valid's binary_logloss: 0.660321
[300] train's binary_logloss: 0.654518 valid's binary_logloss: 0.655319
[400] train's binary_logloss: 0.651377 valid's binary_logloss: 0.652506
[500] train's binary_logloss: 0.649375 valid's binary_logloss: 0.650671
[600] train's binary_logloss: 0.648029 valid's binary_logloss: 0.649863
[700] train's binary_logloss: 0.647077 valid's binary_logloss: 0.649248
[800] train's binary_logloss: 0.64636 valid's binary_logloss: 0.649015
[900] train's binary_logloss: 0.645776 valid's binary_logloss: 0.649025
[1000] train's binary_logloss: 0.645301 valid's binary_logloss: 0.649111
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.645301 valid's binary_logloss: 0.649111
num_leaves, val_score: 0.647726: 85%|########5 | 17/20 [00:14<00:03, 1.12s/it][I 2020-09-27 04:38:38,549] Trial 23 finished with value: 0.6491105632200171 and parameters: {'num_leaves': 2}. Best is trial 9 with value: 0.6477259914776615.
num_leaves, val_score: 0.647726: 85%|########5 | 17/20 [00:14<00:03, 1.12s/it][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005133 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.53935 valid's binary_logloss: 0.652897
Early stopping, best iteration is:
[92] train's binary_logloss: 0.546487 valid's binary_logloss: 0.652502
num_leaves, val_score: 0.647726: 90%|######### | 18/20 [00:15<00:01, 1.04it/s][I 2020-09-27 04:38:39,143] Trial 24 finished with value: 0.6525017174749979 and parameters: {'num_leaves': 48}. Best is trial 9 with value: 0.6477259914776615.
num_leaves, val_score: 0.647726: 90%|######### | 18/20 [00:15<00:01, 1.04it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.003834 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.669306 valid's binary_logloss: 0.670416
[200] train's binary_logloss: 0.659732 valid's binary_logloss: 0.660321
[300] train's binary_logloss: 0.654518 valid's binary_logloss: 0.655319
[400] train's binary_logloss: 0.651377 valid's binary_logloss: 0.652506
[500] train's binary_logloss: 0.649375 valid's binary_logloss: 0.650671
[600] train's binary_logloss: 0.648029 valid's binary_logloss: 0.649863
[700] train's binary_logloss: 0.647077 valid's binary_logloss: 0.649248
[800] train's binary_logloss: 0.64636 valid's binary_logloss: 0.649015
[900] train's binary_logloss: 0.645776 valid's binary_logloss: 0.649025
[1000] train's binary_logloss: 0.645301 valid's binary_logloss: 0.649111
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.645301 valid's binary_logloss: 0.649111
num_leaves, val_score: 0.647726: 95%|#########5| 19/20 [00:16<00:01, 1.02s/it][I 2020-09-27 04:38:40,291] Trial 25 finished with value: 0.649110563220017 and parameters: {'num_leaves': 2}. Best is trial 9 with value: 0.6477259914776615.
num_leaves, val_score: 0.647726: 95%|#########5| 19/20 [00:16<00:01, 1.02s/it][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000543 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.468108 valid's binary_logloss: 0.654681
Early stopping, best iteration is:
[69] train's binary_logloss: 0.512521 valid's binary_logloss: 0.652052
num_leaves, val_score: 0.647726: 100%|##########| 20/20 [00:17<00:00, 1.05it/s][I 2020-09-27 04:38:41,075] Trial 26 finished with value: 0.652051918071773 and parameters: {'num_leaves': 86}. Best is trial 9 with value: 0.6477259914776615.
num_leaves, val_score: 0.647726: 100%|##########| 20/20 [00:17<00:00, 1.15it/s]
bagging, val_score: 0.647726: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.003535 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.63651 valid's binary_logloss: 0.650977
[200] train's binary_logloss: 0.620185 valid's binary_logloss: 0.652728
Early stopping, best iteration is:
[150] train's binary_logloss: 0.627724 valid's binary_logloss: 0.650131
bagging, val_score: 0.647726: 10%|# | 1/10 [00:00<00:04, 1.99it/s][I 2020-09-27 04:38:41,600] Trial 27 finished with value: 0.650131071613353 and parameters: {'bagging_fraction': 0.4586199697400575, 'bagging_freq': 3}. Best is trial 27 with value: 0.650131071613353.
bagging, val_score: 0.647726: 10%|# | 1/10 [00:00<00:04, 1.99it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000853 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.635382 valid's binary_logloss: 0.651145
[200] train's binary_logloss: 0.61788 valid's binary_logloss: 0.647841
[300] train's binary_logloss: 0.603222 valid's binary_logloss: 0.647911
[400] train's binary_logloss: 0.589945 valid's binary_logloss: 0.647612
Early stopping, best iteration is:
[340] train's binary_logloss: 0.597917 valid's binary_logloss: 0.647006
bagging, val_score: 0.647006: 20%|## | 2/10 [00:01<00:05, 1.33it/s][I 2020-09-27 04:38:42,919] Trial 28 finished with value: 0.6470063208679462 and parameters: {'bagging_fraction': 0.9958662267298061, 'bagging_freq': 7}. Best is trial 28 with value: 0.6470063208679462.
bagging, val_score: 0.647006: 20%|## | 2/10 [00:01<00:05, 1.33it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.003784 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.635259 valid's binary_logloss: 0.651142
[200] train's binary_logloss: 0.617565 valid's binary_logloss: 0.64962
Early stopping, best iteration is:
[195] train's binary_logloss: 0.618361 valid's binary_logloss: 0.649343
bagging, val_score: 0.647006: 30%|### | 3/10 [00:02<00:04, 1.45it/s][I 2020-09-27 04:38:43,462] Trial 29 finished with value: 0.6493431744945917 and parameters: {'bagging_fraction': 0.9799916779746923, 'bagging_freq': 7}. Best is trial 28 with value: 0.6470063208679462.
bagging, val_score: 0.647006: 30%|### | 3/10 [00:02<00:04, 1.45it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000779 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.635241 valid's binary_logloss: 0.651269
[200] train's binary_logloss: 0.618036 valid's binary_logloss: 0.649672
[300] train's binary_logloss: 0.603932 valid's binary_logloss: 0.649408
[400] train's binary_logloss: 0.590656 valid's binary_logloss: 0.648685
Early stopping, best iteration is:
[368] train's binary_logloss: 0.594891 valid's binary_logloss: 0.648347
bagging, val_score: 0.647006: 40%|#### | 4/10 [00:03<00:04, 1.37it/s][I 2020-09-27 04:38:44,294] Trial 30 finished with value: 0.6483471798474496 and parameters: {'bagging_fraction': 0.9986614103818544, 'bagging_freq': 7}. Best is trial 28 with value: 0.6470063208679462.
bagging, val_score: 0.647006: 40%|#### | 4/10 [00:03<00:04, 1.37it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007385 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.63573 valid's binary_logloss: 0.649902
[200] train's binary_logloss: 0.618646 valid's binary_logloss: 0.650492
Early stopping, best iteration is:
[170] train's binary_logloss: 0.623387 valid's binary_logloss: 0.648591
bagging, val_score: 0.647006: 50%|##### | 5/10 [00:03<00:03, 1.52it/s][I 2020-09-27 04:38:44,781] Trial 31 finished with value: 0.6485905878341305 and parameters: {'bagging_fraction': 0.7328660882461872, 'bagging_freq': 5}. Best is trial 28 with value: 0.6470063208679462.
bagging, val_score: 0.647006: 50%|##### | 5/10 [00:03<00:03, 1.52it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000825 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.635272 valid's binary_logloss: 0.650422
[200] train's binary_logloss: 0.61747 valid's binary_logloss: 0.648941
[300] train's binary_logloss: 0.602462 valid's binary_logloss: 0.650582
Early stopping, best iteration is:
[227] train's binary_logloss: 0.613139 valid's binary_logloss: 0.648451
bagging, val_score: 0.647006: 60%|###### | 6/10 [00:04<00:02, 1.49it/s][I 2020-09-27 04:38:45,485] Trial 32 finished with value: 0.6484510872456424 and parameters: {'bagging_fraction': 0.7900181913598195, 'bagging_freq': 1}. Best is trial 28 with value: 0.6470063208679462.
bagging, val_score: 0.647006: 60%|###### | 6/10 [00:04<00:02, 1.49it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.003362 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.636504 valid's binary_logloss: 0.650927
[200] train's binary_logloss: 0.620547 valid's binary_logloss: 0.651909
Early stopping, best iteration is:
[115] train's binary_logloss: 0.633633 valid's binary_logloss: 0.6507
bagging, val_score: 0.647006: 70%|####### | 7/10 [00:05<00:02, 1.40it/s][I 2020-09-27 04:38:46,298] Trial 33 finished with value: 0.6506995210705391 and parameters: {'bagging_fraction': 0.47267376312879716, 'bagging_freq': 5}. Best is trial 28 with value: 0.6470063208679462.
bagging, val_score: 0.647006: 70%|####### | 7/10 [00:05<00:02, 1.40it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005528 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.635448 valid's binary_logloss: 0.651834
[200] train's binary_logloss: 0.617764 valid's binary_logloss: 0.650603
[300] train's binary_logloss: 0.603475 valid's binary_logloss: 0.652326
Early stopping, best iteration is:
[226] train's binary_logloss: 0.614069 valid's binary_logloss: 0.649886
bagging, val_score: 0.647006: 80%|######## | 8/10 [00:05<00:01, 1.47it/s][I 2020-09-27 04:38:46,894] Trial 34 finished with value: 0.649885537943031 and parameters: {'bagging_fraction': 0.876875549392964, 'bagging_freq': 7}. Best is trial 28 with value: 0.6470063208679462.
bagging, val_score: 0.647006: 80%|######## | 8/10 [00:05<00:01, 1.47it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000544 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.635851 valid's binary_logloss: 0.649962
[200] train's binary_logloss: 0.6186 valid's binary_logloss: 0.647048
[300] train's binary_logloss: 0.603846 valid's binary_logloss: 0.649943
Early stopping, best iteration is:
[207] train's binary_logloss: 0.617528 valid's binary_logloss: 0.646928
bagging, val_score: 0.646928: 90%|######### | 9/10 [00:06<00:00, 1.56it/s][I 2020-09-27 04:38:47,444] Trial 35 finished with value: 0.6469279610503994 and parameters: {'bagging_fraction': 0.6178680364111143, 'bagging_freq': 3}. Best is trial 35 with value: 0.6469279610503994.
bagging, val_score: 0.646928: 90%|######### | 9/10 [00:06<00:00, 1.56it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000490 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.635909 valid's binary_logloss: 0.648293
[200] train's binary_logloss: 0.618922 valid's binary_logloss: 0.649973
Early stopping, best iteration is:
[149] train's binary_logloss: 0.627266 valid's binary_logloss: 0.647531
bagging, val_score: 0.646928: 100%|##########| 10/10 [00:06<00:00, 1.68it/s][I 2020-09-27 04:38:47,929] Trial 36 finished with value: 0.6475307026936266 and parameters: {'bagging_fraction': 0.5956628615183736, 'bagging_freq': 2}. Best is trial 35 with value: 0.6469279610503994.
bagging, val_score: 0.646928: 100%|##########| 10/10 [00:06<00:00, 1.46it/s]
feature_fraction_stage2, val_score: 0.646928: 0%| | 0/6 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000428 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.634952 valid's binary_logloss: 0.65056
[200] train's binary_logloss: 0.617682 valid's binary_logloss: 0.649553
[300] train's binary_logloss: 0.603161 valid's binary_logloss: 0.652262
Early stopping, best iteration is:
[241] train's binary_logloss: 0.611549 valid's binary_logloss: 0.649321
feature_fraction_stage2, val_score: 0.646928: 17%|#6 | 1/6 [00:00<00:03, 1.60it/s][I 2020-09-27 04:38:48,575] Trial 37 finished with value: 0.6493214800336689 and parameters: {'feature_fraction': 0.748}. Best is trial 37 with value: 0.6493214800336689.
feature_fraction_stage2, val_score: 0.646928: 17%|#6 | 1/6 [00:00<00:03, 1.60it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001333 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.635533 valid's binary_logloss: 0.650663
[200] train's binary_logloss: 0.618659 valid's binary_logloss: 0.651432
Early stopping, best iteration is:
[141] train's binary_logloss: 0.628172 valid's binary_logloss: 0.650017
feature_fraction_stage2, val_score: 0.646928: 33%|###3 | 2/6 [00:01<00:02, 1.67it/s][I 2020-09-27 04:38:49,111] Trial 38 finished with value: 0.6500171564584806 and parameters: {'feature_fraction': 0.652}. Best is trial 37 with value: 0.6493214800336689.
feature_fraction_stage2, val_score: 0.646928: 33%|###3 | 2/6 [00:01<00:02, 1.67it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000674 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.635851 valid's binary_logloss: 0.649962
[200] train's binary_logloss: 0.6186 valid's binary_logloss: 0.647048
[300] train's binary_logloss: 0.603846 valid's binary_logloss: 0.649943
Early stopping, best iteration is:
[207] train's binary_logloss: 0.617528 valid's binary_logloss: 0.646928
feature_fraction_stage2, val_score: 0.646928: 50%|##### | 3/6 [00:02<00:02, 1.33it/s][I 2020-09-27 04:38:50,213] Trial 39 finished with value: 0.6469279610503994 and parameters: {'feature_fraction': 0.6839999999999999}. Best is trial 39 with value: 0.6469279610503994.
feature_fraction_stage2, val_score: 0.646928: 50%|##### | 3/6 [00:02<00:02, 1.33it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000915 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.635444 valid's binary_logloss: 0.649405
[200] train's binary_logloss: 0.618068 valid's binary_logloss: 0.648194
Early stopping, best iteration is:
[173] train's binary_logloss: 0.622667 valid's binary_logloss: 0.647565
feature_fraction_stage2, val_score: 0.646928: 67%|######6 | 4/6 [00:02<00:01, 1.46it/s][I 2020-09-27 04:38:50,749] Trial 40 finished with value: 0.6475654368255805 and parameters: {'feature_fraction': 0.7799999999999999}. Best is trial 39 with value: 0.6469279610503994.
feature_fraction_stage2, val_score: 0.646928: 67%|######6 | 4/6 [00:02<00:01, 1.46it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004998 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.63624 valid's binary_logloss: 0.650875
[200] train's binary_logloss: 0.619436 valid's binary_logloss: 0.649352
Early stopping, best iteration is:
[165] train's binary_logloss: 0.62485 valid's binary_logloss: 0.648991
feature_fraction_stage2, val_score: 0.646928: 83%|########3 | 5/6 [00:03<00:00, 1.59it/s][I 2020-09-27 04:38:51,248] Trial 41 finished with value: 0.6489914374612618 and parameters: {'feature_fraction': 0.62}. Best is trial 39 with value: 0.6469279610503994.
feature_fraction_stage2, val_score: 0.646928: 83%|########3 | 5/6 [00:03<00:00, 1.59it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000866 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.634952 valid's binary_logloss: 0.65056
[200] train's binary_logloss: 0.617682 valid's binary_logloss: 0.649553
[300] train's binary_logloss: 0.603161 valid's binary_logloss: 0.652262
Early stopping, best iteration is:
[241] train's binary_logloss: 0.611549 valid's binary_logloss: 0.649321
feature_fraction_stage2, val_score: 0.646928: 100%|##########| 6/6 [00:03<00:00, 1.56it/s][I 2020-09-27 04:38:51,915] Trial 42 finished with value: 0.6493214800336689 and parameters: {'feature_fraction': 0.716}. Best is trial 39 with value: 0.6469279610503994.
feature_fraction_stage2, val_score: 0.646928: 100%|##########| 6/6 [00:03<00:00, 1.51it/s]
regularization_factors, val_score: 0.646928: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000644 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.635945 valid's binary_logloss: 0.65002
[200] train's binary_logloss: 0.61936 valid's binary_logloss: 0.648569
Early stopping, best iteration is:
[161] train's binary_logloss: 0.625243 valid's binary_logloss: 0.647758
regularization_factors, val_score: 0.646928: 5%|5 | 1/20 [00:00<00:10, 1.73it/s][I 2020-09-27 04:38:52,510] Trial 43 finished with value: 0.6477575542610232 and parameters: {'lambda_l1': 0.5305532619182826, 'lambda_l2': 2.2940234848178292e-05}. Best is trial 43 with value: 0.6477575542610232.
regularization_factors, val_score: 0.646928: 5%|5 | 1/20 [00:00<00:10, 1.73it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004925 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.637159 valid's binary_logloss: 0.648561
[200] train's binary_logloss: 0.62222 valid's binary_logloss: 0.646184
[300] train's binary_logloss: 0.610053 valid's binary_logloss: 0.646865
Early stopping, best iteration is:
[250] train's binary_logloss: 0.616221 valid's binary_logloss: 0.645506
regularization_factors, val_score: 0.645506: 10%|# | 2/20 [00:01<00:12, 1.48it/s][I 2020-09-27 04:38:53,419] Trial 44 finished with value: 0.6455057317841788 and parameters: {'lambda_l1': 1.5729907152074253e-08, 'lambda_l2': 6.408315304180313}. Best is trial 44 with value: 0.6455057317841788.
regularization_factors, val_score: 0.645506: 10%|# | 2/20 [00:01<00:12, 1.48it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010288 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.637347 valid's binary_logloss: 0.64928
[200] train's binary_logloss: 0.623416 valid's binary_logloss: 0.647003
[300] train's binary_logloss: 0.611546 valid's binary_logloss: 0.648015
Early stopping, best iteration is:
[225] train's binary_logloss: 0.620614 valid's binary_logloss: 0.646489
regularization_factors, val_score: 0.645506: 15%|#5 | 3/20 [00:02<00:12, 1.34it/s][I 2020-09-27 04:38:54,330] Trial 45 finished with value: 0.6464889390748203 and parameters: {'lambda_l1': 1.792278679780671e-08, 'lambda_l2': 8.810233526119564}. Best is trial 44 with value: 0.6455057317841788.
regularization_factors, val_score: 0.645506: 15%|#5 | 3/20 [00:02<00:12, 1.34it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000443 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.63666 valid's binary_logloss: 0.649464
[200] train's binary_logloss: 0.62174 valid's binary_logloss: 0.647351
[300] train's binary_logloss: 0.609288 valid's binary_logloss: 0.648443
Early stopping, best iteration is:
[254] train's binary_logloss: 0.615011 valid's binary_logloss: 0.646435
regularization_factors, val_score: 0.645506: 20%|## | 4/20 [00:03<00:11, 1.37it/s][I 2020-09-27 04:38:55,027] Trial 46 finished with value: 0.6464351970725228 and parameters: {'lambda_l1': 2.2856982982765707e-08, 'lambda_l2': 4.5894526330871095}. Best is trial 44 with value: 0.6455057317841788.
regularization_factors, val_score: 0.645506: 20%|## | 4/20 [00:03<00:11, 1.37it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.007799 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.637474 valid's binary_logloss: 0.648583
[200] train's binary_logloss: 0.62337 valid's binary_logloss: 0.645859
[300] train's binary_logloss: 0.61123 valid's binary_logloss: 0.645293
[400] train's binary_logloss: 0.599843 valid's binary_logloss: 0.646225
Early stopping, best iteration is:
[317] train's binary_logloss: 0.609063 valid's binary_logloss: 0.644762
regularization_factors, val_score: 0.644762: 25%|##5 | 5/20 [00:03<00:11, 1.32it/s][I 2020-09-27 04:38:55,839] Trial 47 finished with value: 0.6447624933131433 and parameters: {'lambda_l1': 1.2755111716437187e-08, 'lambda_l2': 9.850714874340618}. Best is trial 47 with value: 0.6447624933131433.
regularization_factors, val_score: 0.644762: 25%|##5 | 5/20 [00:03<00:11, 1.32it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004054 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.63754 valid's binary_logloss: 0.648978
[200] train's binary_logloss: 0.623228 valid's binary_logloss: 0.646323
[300] train's binary_logloss: 0.611727 valid's binary_logloss: 0.64615
[400] train's binary_logloss: 0.600473 valid's binary_logloss: 0.646062
Early stopping, best iteration is:
[347] train's binary_logloss: 0.606258 valid's binary_logloss: 0.645207
regularization_factors, val_score: 0.644762: 30%|### | 6/20 [00:04<00:11, 1.26it/s][I 2020-09-27 04:38:56,723] Trial 48 finished with value: 0.6452071554754294 and parameters: {'lambda_l1': 1.3011629397041354e-08, 'lambda_l2': 9.988192496777456}. Best is trial 47 with value: 0.6447624933131433.
regularization_factors, val_score: 0.644762: 30%|### | 6/20 [00:04<00:11, 1.26it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007536 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.6371 valid's binary_logloss: 0.6496
[200] train's binary_logloss: 0.622819 valid's binary_logloss: 0.647511
[300] train's binary_logloss: 0.610769 valid's binary_logloss: 0.647971
Early stopping, best iteration is:
[232] train's binary_logloss: 0.618821 valid's binary_logloss: 0.646619
regularization_factors, val_score: 0.644762: 35%|###5 | 7/20 [00:05<00:11, 1.14it/s][I 2020-09-27 04:38:57,789] Trial 49 finished with value: 0.6466188865283253 and parameters: {'lambda_l1': 1.5462932963049097e-08, 'lambda_l2': 8.71269927175754}. Best is trial 47 with value: 0.6447624933131433.
regularization_factors, val_score: 0.644762: 35%|###5 | 7/20 [00:05<00:11, 1.14it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004606 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.637133 valid's binary_logloss: 0.649344
[200] train's binary_logloss: 0.622572 valid's binary_logloss: 0.647963
[300] train's binary_logloss: 0.61055 valid's binary_logloss: 0.647491
Early stopping, best iteration is:
[255] train's binary_logloss: 0.616042 valid's binary_logloss: 0.646564
regularization_factors, val_score: 0.644762: 40%|#### | 8/20 [00:06<00:09, 1.21it/s][I 2020-09-27 04:38:58,493] Trial 50 finished with value: 0.6465642008648574 and parameters: {'lambda_l1': 1.1089576217805702e-08, 'lambda_l2': 7.312599056585437}. Best is trial 47 with value: 0.6447624933131433.
regularization_factors, val_score: 0.644762: 40%|#### | 8/20 [00:06<00:09, 1.21it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000434 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.637479 valid's binary_logloss: 0.648584
[200] train's binary_logloss: 0.62341 valid's binary_logloss: 0.645728
[300] train's binary_logloss: 0.611552 valid's binary_logloss: 0.64603
Early stopping, best iteration is:
[254] train's binary_logloss: 0.617107 valid's binary_logloss: 0.644789
regularization_factors, val_score: 0.644762: 45%|####5 | 9/20 [00:07<00:08, 1.27it/s][I 2020-09-27 04:38:59,196] Trial 51 finished with value: 0.6447887069638869 and parameters: {'lambda_l1': 1.7016389032742813e-08, 'lambda_l2': 9.898211761649065}. Best is trial 47 with value: 0.6447624933131433.
regularization_factors, val_score: 0.644762: 45%|####5 | 9/20 [00:07<00:08, 1.27it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007899 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.637278 valid's binary_logloss: 0.649089
[200] train's binary_logloss: 0.622576 valid's binary_logloss: 0.647208
[300] train's binary_logloss: 0.610418 valid's binary_logloss: 0.648157
Early stopping, best iteration is:
[255] train's binary_logloss: 0.615802 valid's binary_logloss: 0.646398
regularization_factors, val_score: 0.644762: 50%|##### | 10/20 [00:07<00:07, 1.31it/s][I 2020-09-27 04:38:59,894] Trial 52 finished with value: 0.6463976805891822 and parameters: {'lambda_l1': 2.2240739211030176e-08, 'lambda_l2': 7.404538655662823}. Best is trial 47 with value: 0.6447624933131433.
regularization_factors, val_score: 0.644762: 50%|##### | 10/20 [00:07<00:07, 1.31it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000572 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.637182 valid's binary_logloss: 0.647682
[200] train's binary_logloss: 0.623009 valid's binary_logloss: 0.646084
[300] train's binary_logloss: 0.611035 valid's binary_logloss: 0.645152
Early stopping, best iteration is:
[260] train's binary_logloss: 0.615822 valid's binary_logloss: 0.643798
regularization_factors, val_score: 0.643798: 55%|#####5 | 11/20 [00:08<00:06, 1.35it/s][I 2020-09-27 04:39:00,581] Trial 53 finished with value: 0.6437977363337001 and parameters: {'lambda_l1': 2.039143182350295e-08, 'lambda_l2': 9.400199053113356}. Best is trial 53 with value: 0.6437977363337001.
regularization_factors, val_score: 0.643798: 55%|#####5 | 11/20 [00:08<00:06, 1.35it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004032 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.637206 valid's binary_logloss: 0.647684
[200] train's binary_logloss: 0.623151 valid's binary_logloss: 0.646089
[300] train's binary_logloss: 0.611403 valid's binary_logloss: 0.645967
Early stopping, best iteration is:
[255] train's binary_logloss: 0.616582 valid's binary_logloss: 0.644531
regularization_factors, val_score: 0.643798: 60%|###### | 12/20 [00:09<00:06, 1.17it/s][I 2020-09-27 04:39:01,697] Trial 54 finished with value: 0.6445311936163728 and parameters: {'lambda_l1': 1.6270643281831753e-08, 'lambda_l2': 9.628307902916124}. Best is trial 53 with value: 0.6437977363337001.
regularization_factors, val_score: 0.643798: 60%|###### | 12/20 [00:09<00:06, 1.17it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000874 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.637281 valid's binary_logloss: 0.649044
[200] train's binary_logloss: 0.622584 valid's binary_logloss: 0.647021
[300] train's binary_logloss: 0.610763 valid's binary_logloss: 0.647525
Early stopping, best iteration is:
[254] train's binary_logloss: 0.616357 valid's binary_logloss: 0.645998
regularization_factors, val_score: 0.643798: 65%|######5 | 13/20 [00:10<00:05, 1.22it/s][I 2020-09-27 04:39:02,449] Trial 55 finished with value: 0.6459984456397755 and parameters: {'lambda_l1': 1.0240697224966498e-08, 'lambda_l2': 8.012028500403863}. Best is trial 53 with value: 0.6437977363337001.
regularization_factors, val_score: 0.643798: 65%|######5 | 13/20 [00:10<00:05, 1.22it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000558 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.635858 valid's binary_logloss: 0.649961
[200] train's binary_logloss: 0.618625 valid's binary_logloss: 0.647047
[300] train's binary_logloss: 0.603932 valid's binary_logloss: 0.648748
Early stopping, best iteration is:
[207] train's binary_logloss: 0.617554 valid's binary_logloss: 0.646927
regularization_factors, val_score: 0.643798: 70%|####### | 14/20 [00:11<00:04, 1.32it/s][I 2020-09-27 04:39:03,059] Trial 56 finished with value: 0.6469267717887545 and parameters: {'lambda_l1': 1.2959832015054276e-06, 'lambda_l2': 0.01861979600536283}. Best is trial 53 with value: 0.6437977363337001.
regularization_factors, val_score: 0.643798: 70%|####### | 14/20 [00:11<00:04, 1.32it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010899 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.635708 valid's binary_logloss: 0.649563
[200] train's binary_logloss: 0.61871 valid's binary_logloss: 0.647652
[300] train's binary_logloss: 0.604402 valid's binary_logloss: 0.650241
Early stopping, best iteration is:
[206] train's binary_logloss: 0.6179 valid's binary_logloss: 0.647417
regularization_factors, val_score: 0.643798: 75%|#######5 | 15/20 [00:11<00:03, 1.43it/s][I 2020-09-27 04:39:03,628] Trial 57 finished with value: 0.6474166124124395 and parameters: {'lambda_l1': 5.765532893767189e-07, 'lambda_l2': 0.14788147344128014}. Best is trial 53 with value: 0.6437977363337001.
regularization_factors, val_score: 0.643798: 75%|#######5 | 15/20 [00:11<00:03, 1.43it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005282 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.635911 valid's binary_logloss: 0.650092
[200] train's binary_logloss: 0.618697 valid's binary_logloss: 0.649148
[300] train's binary_logloss: 0.604075 valid's binary_logloss: 0.650674
Early stopping, best iteration is:
[250] train's binary_logloss: 0.611662 valid's binary_logloss: 0.648499
regularization_factors, val_score: 0.643798: 80%|######## | 16/20 [00:12<00:02, 1.44it/s][I 2020-09-27 04:39:04,302] Trial 58 finished with value: 0.6484991507925789 and parameters: {'lambda_l1': 5.203220654613759e-07, 'lambda_l2': 0.30810693370096226}. Best is trial 53 with value: 0.6437977363337001.
regularization_factors, val_score: 0.643798: 80%|######## | 16/20 [00:12<00:02, 1.44it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004372 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.63618 valid's binary_logloss: 0.650527
[200] train's binary_logloss: 0.619801 valid's binary_logloss: 0.649123
Early stopping, best iteration is:
[160] train's binary_logloss: 0.625979 valid's binary_logloss: 0.648367
regularization_factors, val_score: 0.643798: 85%|########5 | 17/20 [00:13<00:02, 1.37it/s][I 2020-09-27 04:39:05,120] Trial 59 finished with value: 0.6483666505535219 and parameters: {'lambda_l1': 1.1681262794154124e-07, 'lambda_l2': 0.9991529121249367}. Best is trial 53 with value: 0.6437977363337001.
regularization_factors, val_score: 0.643798: 85%|########5 | 17/20 [00:13<00:02, 1.37it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008489 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.635851 valid's binary_logloss: 0.649962
[200] train's binary_logloss: 0.6186 valid's binary_logloss: 0.647048
[300] train's binary_logloss: 0.603848 valid's binary_logloss: 0.649943
Early stopping, best iteration is:
[207] train's binary_logloss: 0.617528 valid's binary_logloss: 0.646928
regularization_factors, val_score: 0.643798: 90%|######### | 18/20 [00:13<00:01, 1.40it/s][I 2020-09-27 04:39:05,803] Trial 60 finished with value: 0.6469279610385475 and parameters: {'lambda_l1': 1.213400855943739e-08, 'lambda_l2': 1.534183918800275e-07}. Best is trial 53 with value: 0.6437977363337001.
regularization_factors, val_score: 0.643798: 90%|######### | 18/20 [00:13<00:01, 1.40it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009995 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.637219 valid's binary_logloss: 0.650017
[200] train's binary_logloss: 0.622187 valid's binary_logloss: 0.647211
[300] train's binary_logloss: 0.609711 valid's binary_logloss: 0.647263
Early stopping, best iteration is:
[254] train's binary_logloss: 0.615481 valid's binary_logloss: 0.64613
regularization_factors, val_score: 0.643798: 95%|#########5| 19/20 [00:14<00:00, 1.44it/s][I 2020-09-27 04:39:06,446] Trial 61 finished with value: 0.6461303865450015 and parameters: {'lambda_l1': 1.0232750932997174e-08, 'lambda_l2': 5.61104076150738}. Best is trial 53 with value: 0.6437977363337001.
regularization_factors, val_score: 0.643798: 95%|#########5| 19/20 [00:14<00:00, 1.44it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000915 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.637448 valid's binary_logloss: 0.649065
[200] train's binary_logloss: 0.623392 valid's binary_logloss: 0.647489
[300] train's binary_logloss: 0.611817 valid's binary_logloss: 0.647888
Early stopping, best iteration is:
[236] train's binary_logloss: 0.61913 valid's binary_logloss: 0.647044
regularization_factors, val_score: 0.643798: 100%|##########| 20/20 [00:15<00:00, 1.48it/s][I 2020-09-27 04:39:07,086] Trial 62 finished with value: 0.6470438910627877 and parameters: {'lambda_l1': 1.4832304242797658e-08, 'lambda_l2': 9.997088293767975}. Best is trial 53 with value: 0.6437977363337001.
regularization_factors, val_score: 0.643798: 100%|##########| 20/20 [00:15<00:00, 1.32it/s]
min_data_in_leaf, val_score: 0.643798: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.003521 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.637289 valid's binary_logloss: 0.648183
[200] train's binary_logloss: 0.62309 valid's binary_logloss: 0.64706
[300] train's binary_logloss: 0.611616 valid's binary_logloss: 0.64692
Early stopping, best iteration is:
[254] train's binary_logloss: 0.616891 valid's binary_logloss: 0.645943
min_data_in_leaf, val_score: 0.643798: 20%|## | 1/5 [00:00<00:02, 1.35it/s][I 2020-09-27 04:39:07,843] Trial 63 finished with value: 0.6459430156769583 and parameters: {'min_child_samples': 50}. Best is trial 63 with value: 0.6459430156769583.
min_data_in_leaf, val_score: 0.643798: 20%|## | 1/5 [00:00<00:02, 1.35it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000832 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.637182 valid's binary_logloss: 0.647682
[200] train's binary_logloss: 0.622869 valid's binary_logloss: 0.646623
[300] train's binary_logloss: 0.611124 valid's binary_logloss: 0.646202
Early stopping, best iteration is:
[254] train's binary_logloss: 0.616639 valid's binary_logloss: 0.645162
min_data_in_leaf, val_score: 0.643798: 40%|#### | 2/5 [00:01<00:02, 1.27it/s][I 2020-09-27 04:39:08,731] Trial 64 finished with value: 0.6451618189346408 and parameters: {'min_child_samples': 10}. Best is trial 64 with value: 0.6451618189346408.
min_data_in_leaf, val_score: 0.643798: 40%|#### | 2/5 [00:01<00:02, 1.27it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012841 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.637182 valid's binary_logloss: 0.647682
[200] train's binary_logloss: 0.623074 valid's binary_logloss: 0.64581
[300] train's binary_logloss: 0.61124 valid's binary_logloss: 0.645222
Early stopping, best iteration is:
[255] train's binary_logloss: 0.616649 valid's binary_logloss: 0.643974
min_data_in_leaf, val_score: 0.643798: 60%|###### | 3/5 [00:02<00:01, 1.11it/s][I 2020-09-27 04:39:09,890] Trial 65 finished with value: 0.6439740196693694 and parameters: {'min_child_samples': 25}. Best is trial 65 with value: 0.6439740196693694.
min_data_in_leaf, val_score: 0.643798: 60%|###### | 3/5 [00:02<00:01, 1.11it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004742 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.637182 valid's binary_logloss: 0.647682
[200] train's binary_logloss: 0.622869 valid's binary_logloss: 0.646623
[300] train's binary_logloss: 0.610951 valid's binary_logloss: 0.646227
Early stopping, best iteration is:
[243] train's binary_logloss: 0.617796 valid's binary_logloss: 0.645499
min_data_in_leaf, val_score: 0.643798: 80%|######## | 4/5 [00:03<00:00, 1.19it/s][I 2020-09-27 04:39:10,600] Trial 66 finished with value: 0.6454986987713198 and parameters: {'min_child_samples': 5}. Best is trial 65 with value: 0.6439740196693694.
min_data_in_leaf, val_score: 0.643798: 80%|######## | 4/5 [00:03<00:00, 1.19it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000441 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695
[LightGBM] [Info] Start training from score 0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.637729 valid's binary_logloss: 0.649258
[200] train's binary_logloss: 0.62325 valid's binary_logloss: 0.646308
[300] train's binary_logloss: 0.611957 valid's binary_logloss: 0.647641
Early stopping, best iteration is:
[238] train's binary_logloss: 0.618817 valid's binary_logloss: 0.645531
min_data_in_leaf, val_score: 0.643798: 100%|##########| 5/5 [00:04<00:00, 1.26it/s][I 2020-09-27 04:39:11,288] Trial 67 finished with value: 0.6455312591688435 and parameters: {'min_child_samples': 100}. Best is trial 65 with value: 0.6439740196693694.
min_data_in_leaf, val_score: 0.643798: 100%|##########| 5/5 [00:04<00:00, 1.19it/s]
Fold : 1
[I 2020-09-27 04:39:11,344] A new study created in memory with name: no-name-8ec748bd-7c24-485d-8dfe-5099f051a9d9
feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001101 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.575663 valid's binary_logloss: 0.661357
Early stopping, best iteration is:
[56] train's binary_logloss: 0.605835 valid's binary_logloss: 0.659163
feature_fraction, val_score: 0.659163: 14%|#4 | 1/7 [00:00<00:03, 1.89it/s][I 2020-09-27 04:39:11,890] Trial 0 finished with value: 0.6591626166519308 and parameters: {'feature_fraction': 0.7}. Best is trial 0 with value: 0.6591626166519308.
feature_fraction, val_score: 0.659163: 14%|#4 | 1/7 [00:00<00:03, 1.89it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000533 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.577716 valid's binary_logloss: 0.658007
Early stopping, best iteration is:
[72] train's binary_logloss: 0.596211 valid's binary_logloss: 0.65673
feature_fraction, val_score: 0.656730: 29%|##8 | 2/7 [00:00<00:02, 1.98it/s][I 2020-09-27 04:39:12,340] Trial 1 finished with value: 0.6567303283064773 and parameters: {'feature_fraction': 0.5}. Best is trial 1 with value: 0.6567303283064773.
feature_fraction, val_score: 0.656730: 29%|##8 | 2/7 [00:00<00:02, 1.98it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001310 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.576271 valid's binary_logloss: 0.660374
Early stopping, best iteration is:
[66] train's binary_logloss: 0.598062 valid's binary_logloss: 0.658574
feature_fraction, val_score: 0.656730: 43%|####2 | 3/7 [00:01<00:02, 1.53it/s][I 2020-09-27 04:39:13,335] Trial 2 finished with value: 0.6585740014120882 and parameters: {'feature_fraction': 0.6}. Best is trial 1 with value: 0.6567303283064773.
feature_fraction, val_score: 0.656730: 43%|####2 | 3/7 [00:01<00:02, 1.53it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000370 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.582075 valid's binary_logloss: 0.6565
Early stopping, best iteration is:
[94] train's binary_logloss: 0.58569 valid's binary_logloss: 0.655992
feature_fraction, val_score: 0.655992: 57%|#####7 | 4/7 [00:02<00:01, 1.67it/s][I 2020-09-27 04:39:13,802] Trial 3 finished with value: 0.6559924908506093 and parameters: {'feature_fraction': 0.4}. Best is trial 3 with value: 0.6559924908506093.
feature_fraction, val_score: 0.655992: 57%|#####7 | 4/7 [00:02<00:01, 1.67it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000609 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.571094 valid's binary_logloss: 0.661412
Early stopping, best iteration is:
[54] train's binary_logloss: 0.604656 valid's binary_logloss: 0.657544
feature_fraction, val_score: 0.655992: 71%|#######1 | 5/7 [00:02<00:01, 1.78it/s][I 2020-09-27 04:39:14,285] Trial 4 finished with value: 0.6575438872927727 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 3 with value: 0.6559924908506093.
feature_fraction, val_score: 0.655992: 71%|#######1 | 5/7 [00:02<00:01, 1.78it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000923 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.569935 valid's binary_logloss: 0.659342
Early stopping, best iteration is:
[54] train's binary_logloss: 0.602845 valid's binary_logloss: 0.658926
feature_fraction, val_score: 0.655992: 86%|########5 | 6/7 [00:03<00:00, 1.86it/s][I 2020-09-27 04:39:14,768] Trial 5 finished with value: 0.658926225620644 and parameters: {'feature_fraction': 1.0}. Best is trial 3 with value: 0.6559924908506093.
feature_fraction, val_score: 0.655992: 86%|########5 | 6/7 [00:03<00:00, 1.86it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000465 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.574321 valid's binary_logloss: 0.659912
Early stopping, best iteration is:
[66] train's binary_logloss: 0.595979 valid's binary_logloss: 0.657826
feature_fraction, val_score: 0.655992: 100%|##########| 7/7 [00:03<00:00, 1.91it/s][I 2020-09-27 04:39:15,259] Trial 6 finished with value: 0.65782592600512 and parameters: {'feature_fraction': 0.8}. Best is trial 3 with value: 0.6559924908506093.
feature_fraction, val_score: 0.655992: 100%|##########| 7/7 [00:03<00:00, 1.79it/s]
num_leaves, val_score: 0.655992: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000434 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.631152 valid's binary_logloss: 0.657113
[200] train's binary_logloss: 0.611555 valid's binary_logloss: 0.657151
Early stopping, best iteration is:
[114] train's binary_logloss: 0.627933 valid's binary_logloss: 0.656291
num_leaves, val_score: 0.655992: 5%|5 | 1/20 [00:00<00:07, 2.47it/s][I 2020-09-27 04:39:15,682] Trial 7 finished with value: 0.6562905644151281 and parameters: {'num_leaves': 10}. Best is trial 7 with value: 0.6562905644151281.
num_leaves, val_score: 0.655992: 5%|5 | 1/20 [00:00<00:07, 2.47it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000313 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.313854 valid's binary_logloss: 0.672207
Early stopping, best iteration is:
[34] train's binary_logloss: 0.498751 valid's binary_logloss: 0.663462
num_leaves, val_score: 0.655992: 10%|# | 2/20 [00:01<00:13, 1.32it/s][I 2020-09-27 04:39:17,270] Trial 8 finished with value: 0.6634620451568378 and parameters: {'num_leaves': 238}. Best is trial 7 with value: 0.6562905644151281.
num_leaves, val_score: 0.655992: 10%|# | 2/20 [00:02<00:13, 1.32it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000229 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.442275 valid's binary_logloss: 0.665841
Early stopping, best iteration is:
[40] train's binary_logloss: 0.552251 valid's binary_logloss: 0.660374
num_leaves, val_score: 0.655992: 15%|#5 | 3/20 [00:02<00:12, 1.37it/s][I 2020-09-27 04:39:17,923] Trial 9 finished with value: 0.6603744399138358 and parameters: {'num_leaves': 116}. Best is trial 7 with value: 0.6562905644151281.
num_leaves, val_score: 0.655992: 15%|#5 | 3/20 [00:02<00:12, 1.37it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010332 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.308823 valid's binary_logloss: 0.670809
Early stopping, best iteration is:
[36] train's binary_logloss: 0.48791 valid's binary_logloss: 0.665738
num_leaves, val_score: 0.655992: 20%|## | 4/20 [00:03<00:12, 1.27it/s][I 2020-09-27 04:39:18,840] Trial 10 finished with value: 0.6657380834337528 and parameters: {'num_leaves': 244}. Best is trial 7 with value: 0.6562905644151281.
num_leaves, val_score: 0.655992: 20%|## | 4/20 [00:03<00:12, 1.27it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000322 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.658235 valid's binary_logloss: 0.664731
[200] train's binary_logloss: 0.647762 valid's binary_logloss: 0.657647
[300] train's binary_logloss: 0.642542 valid's binary_logloss: 0.655465
[400] train's binary_logloss: 0.638941 valid's binary_logloss: 0.654486
[500] train's binary_logloss: 0.636035 valid's binary_logloss: 0.653981
[600] train's binary_logloss: 0.63343 valid's binary_logloss: 0.65371
Early stopping, best iteration is:
[555] train's binary_logloss: 0.634597 valid's binary_logloss: 0.653553
num_leaves, val_score: 0.653553: 25%|##5 | 5/20 [00:04<00:11, 1.26it/s][I 2020-09-27 04:39:19,645] Trial 11 finished with value: 0.6535525819216924 and parameters: {'num_leaves': 3}. Best is trial 11 with value: 0.6535525819216924.
num_leaves, val_score: 0.653553: 25%|##5 | 5/20 [00:04<00:11, 1.26it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000332 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.620698 valid's binary_logloss: 0.657525
[200] train's binary_logloss: 0.594023 valid's binary_logloss: 0.657636
Early stopping, best iteration is:
[151] train's binary_logloss: 0.606458 valid's binary_logloss: 0.657287
num_leaves, val_score: 0.653553: 30%|### | 6/20 [00:05<00:10, 1.32it/s][I 2020-09-27 04:39:20,325] Trial 12 finished with value: 0.6572874929871441 and parameters: {'num_leaves': 14}. Best is trial 11 with value: 0.6535525819216924.
num_leaves, val_score: 0.653553: 30%|### | 6/20 [00:05<00:10, 1.32it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011876 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.467946 valid's binary_logloss: 0.670734
Early stopping, best iteration is:
[45] train's binary_logloss: 0.556356 valid's binary_logloss: 0.664088
num_leaves, val_score: 0.653553: 35%|###5 | 7/20 [00:05<00:10, 1.25it/s][I 2020-09-27 04:39:21,228] Trial 13 finished with value: 0.6640878662544877 and parameters: {'num_leaves': 97}. Best is trial 11 with value: 0.6535525819216924.
num_leaves, val_score: 0.653553: 35%|###5 | 7/20 [00:05<00:10, 1.25it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000489 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.524406 valid's binary_logloss: 0.663032
Early stopping, best iteration is:
[41] train's binary_logloss: 0.595701 valid's binary_logloss: 0.662191
num_leaves, val_score: 0.653553: 40%|#### | 8/20 [00:06<00:08, 1.39it/s][I 2020-09-27 04:39:21,759] Trial 14 finished with value: 0.6621905066342226 and parameters: {'num_leaves': 62}. Best is trial 11 with value: 0.6535525819216924.
num_leaves, val_score: 0.653553: 40%|#### | 8/20 [00:06<00:08, 1.39it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000238 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.367271 valid's binary_logloss: 0.673756
Early stopping, best iteration is:
[34] train's binary_logloss: 0.528584 valid's binary_logloss: 0.666502
num_leaves, val_score: 0.653553: 45%|####5 | 9/20 [00:07<00:08, 1.30it/s][I 2020-09-27 04:39:22,638] Trial 15 finished with value: 0.6665016344232115 and parameters: {'num_leaves': 180}. Best is trial 11 with value: 0.6535525819216924.
num_leaves, val_score: 0.653553: 45%|####5 | 9/20 [00:07<00:08, 1.30it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000232 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.36627 valid's binary_logloss: 0.674896
Early stopping, best iteration is:
[36] train's binary_logloss: 0.521716 valid's binary_logloss: 0.66343
num_leaves, val_score: 0.653553: 50%|##### | 10/20 [00:08<00:08, 1.21it/s][I 2020-09-27 04:39:23,590] Trial 16 finished with value: 0.6634300540848975 and parameters: {'num_leaves': 183}. Best is trial 11 with value: 0.6535525819216924.
num_leaves, val_score: 0.653553: 50%|##### | 10/20 [00:08<00:08, 1.21it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000415 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.536123 valid's binary_logloss: 0.658196
Early stopping, best iteration is:
[92] train's binary_logloss: 0.543103 valid's binary_logloss: 0.657473
num_leaves, val_score: 0.653553: 55%|#####5 | 11/20 [00:09<00:07, 1.13it/s][I 2020-09-27 04:39:24,613] Trial 17 finished with value: 0.6574725321803361 and parameters: {'num_leaves': 55}. Best is trial 11 with value: 0.6535525819216924.
num_leaves, val_score: 0.653553: 55%|#####5 | 11/20 [00:09<00:07, 1.13it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000386 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.386366 valid's binary_logloss: 0.670881
Early stopping, best iteration is:
[32] train's binary_logloss: 0.54618 valid's binary_logloss: 0.663828
num_leaves, val_score: 0.653553: 60%|###### | 12/20 [00:10<00:07, 1.14it/s][I 2020-09-27 04:39:25,475] Trial 18 finished with value: 0.663827571242285 and parameters: {'num_leaves': 162}. Best is trial 11 with value: 0.6535525819216924.
num_leaves, val_score: 0.653553: 60%|###### | 12/20 [00:10<00:07, 1.14it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000380 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.543127 valid's binary_logloss: 0.660244
Early stopping, best iteration is:
[92] train's binary_logloss: 0.549667 valid's binary_logloss: 0.659503
num_leaves, val_score: 0.653553: 65%|######5 | 13/20 [00:10<00:05, 1.28it/s][I 2020-09-27 04:39:26,025] Trial 19 finished with value: 0.6595031085541457 and parameters: {'num_leaves': 51}. Best is trial 11 with value: 0.6535525819216924.
num_leaves, val_score: 0.653553: 65%|######5 | 13/20 [00:10<00:05, 1.28it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000521 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.636989 valid's binary_logloss: 0.657697
[200] train's binary_logloss: 0.620568 valid's binary_logloss: 0.656088
Early stopping, best iteration is:
[171] train's binary_logloss: 0.624795 valid's binary_logloss: 0.655485
num_leaves, val_score: 0.653553: 70%|####### | 14/20 [00:11<00:04, 1.49it/s][I 2020-09-27 04:39:26,451] Trial 20 finished with value: 0.6554852556407339 and parameters: {'num_leaves': 8}. Best is trial 11 with value: 0.6535525819216924.
num_leaves, val_score: 0.653553: 70%|####### | 14/20 [00:11<00:04, 1.49it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004552 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.643142 valid's binary_logloss: 0.658677
[200] train's binary_logloss: 0.62998 valid's binary_logloss: 0.656522
[300] train's binary_logloss: 0.620354 valid's binary_logloss: 0.655925
Early stopping, best iteration is:
[297] train's binary_logloss: 0.620584 valid's binary_logloss: 0.655776
num_leaves, val_score: 0.653553: 75%|#######5 | 15/20 [00:11<00:03, 1.58it/s][I 2020-09-27 04:39:26,995] Trial 21 finished with value: 0.655776407729715 and parameters: {'num_leaves': 6}. Best is trial 11 with value: 0.6535525819216924.
num_leaves, val_score: 0.653553: 75%|#######5 | 15/20 [00:11<00:03, 1.58it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000245 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.639989 valid's binary_logloss: 0.657478
[200] train's binary_logloss: 0.62545 valid's binary_logloss: 0.65644
Early stopping, best iteration is:
[140] train's binary_logloss: 0.633249 valid's binary_logloss: 0.655769
num_leaves, val_score: 0.653553: 80%|######## | 16/20 [00:12<00:02, 1.77it/s][I 2020-09-27 04:39:27,399] Trial 22 finished with value: 0.6557689832038824 and parameters: {'num_leaves': 7}. Best is trial 11 with value: 0.6535525819216924.
num_leaves, val_score: 0.653553: 80%|######## | 16/20 [00:12<00:02, 1.77it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004511 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.575815 valid's binary_logloss: 0.658024
[200] train's binary_logloss: 0.522157 valid's binary_logloss: 0.657932
Early stopping, best iteration is:
[154] train's binary_logloss: 0.544614 valid's binary_logloss: 0.656365
num_leaves, val_score: 0.653553: 85%|########5 | 17/20 [00:13<00:02, 1.40it/s][I 2020-09-27 04:39:28,462] Trial 23 finished with value: 0.6563652624863247 and parameters: {'num_leaves': 34}. Best is trial 11 with value: 0.6535525819216924.
num_leaves, val_score: 0.653553: 85%|########5 | 17/20 [00:13<00:02, 1.40it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009938 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.647155 valid's binary_logloss: 0.660323
[200] train's binary_logloss: 0.635295 valid's binary_logloss: 0.655924
Early stopping, best iteration is:
[188] train's binary_logloss: 0.63642 valid's binary_logloss: 0.655643
num_leaves, val_score: 0.653553: 90%|######### | 18/20 [00:13<00:01, 1.59it/s][I 2020-09-27 04:39:28,891] Trial 24 finished with value: 0.655643287434117 and parameters: {'num_leaves': 5}. Best is trial 11 with value: 0.6535525819216924.
num_leaves, val_score: 0.653553: 90%|######### | 18/20 [00:13<00:01, 1.59it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000491 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.488408 valid's binary_logloss: 0.664489
Early stopping, best iteration is:
[78] train's binary_logloss: 0.517222 valid's binary_logloss: 0.661867
num_leaves, val_score: 0.653553: 95%|#########5| 19/20 [00:14<00:00, 1.53it/s][I 2020-09-27 04:39:29,600] Trial 25 finished with value: 0.6618666948452248 and parameters: {'num_leaves': 84}. Best is trial 11 with value: 0.6535525819216924.
num_leaves, val_score: 0.653553: 95%|#########5| 19/20 [00:14<00:00, 1.53it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000236 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.580318 valid's binary_logloss: 0.659611
Early stopping, best iteration is:
[67] train's binary_logloss: 0.601799 valid's binary_logloss: 0.658994
num_leaves, val_score: 0.653553: 100%|##########| 20/20 [00:14<00:00, 1.71it/s][I 2020-09-27 04:39:30,031] Trial 26 finished with value: 0.6589942104081719 and parameters: {'num_leaves': 32}. Best is trial 11 with value: 0.6535525819216924.
num_leaves, val_score: 0.653553: 100%|##########| 20/20 [00:14<00:00, 1.35it/s]
bagging, val_score: 0.653553: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.020148 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657525 valid's binary_logloss: 0.662893
[200] train's binary_logloss: 0.647047 valid's binary_logloss: 0.656095
[300] train's binary_logloss: 0.641935 valid's binary_logloss: 0.654303
[400] train's binary_logloss: 0.638294 valid's binary_logloss: 0.653695
[500] train's binary_logloss: 0.635159 valid's binary_logloss: 0.653205
[600] train's binary_logloss: 0.632193 valid's binary_logloss: 0.652803
[700] train's binary_logloss: 0.629395 valid's binary_logloss: 0.651799
[800] train's binary_logloss: 0.626828 valid's binary_logloss: 0.652063
Early stopping, best iteration is:
[745] train's binary_logloss: 0.628188 valid's binary_logloss: 0.651395
bagging, val_score: 0.651395: 10%|# | 1/10 [00:01<00:10, 1.19s/it][I 2020-09-27 04:39:31,244] Trial 27 finished with value: 0.6513952391116833 and parameters: {'bagging_fraction': 0.8097719374771206, 'bagging_freq': 2}. Best is trial 27 with value: 0.6513952391116833.
bagging, val_score: 0.651395: 10%|# | 1/10 [00:01<00:10, 1.19s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000403 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657637 valid's binary_logloss: 0.663101
[200] train's binary_logloss: 0.647154 valid's binary_logloss: 0.656096
[300] train's binary_logloss: 0.642211 valid's binary_logloss: 0.653702
[400] train's binary_logloss: 0.638597 valid's binary_logloss: 0.652862
[500] train's binary_logloss: 0.635489 valid's binary_logloss: 0.652354
[600] train's binary_logloss: 0.632512 valid's binary_logloss: 0.652123
[700] train's binary_logloss: 0.62981 valid's binary_logloss: 0.651852
[800] train's binary_logloss: 0.627226 valid's binary_logloss: 0.651697
Early stopping, best iteration is:
[743] train's binary_logloss: 0.628657 valid's binary_logloss: 0.651255
bagging, val_score: 0.651255: 20%|## | 2/10 [00:02<00:10, 1.35s/it][I 2020-09-27 04:39:32,959] Trial 28 finished with value: 0.6512546217026939 and parameters: {'bagging_fraction': 0.8280339557117807, 'bagging_freq': 2}. Best is trial 28 with value: 0.6512546217026939.
bagging, val_score: 0.651255: 20%|## | 2/10 [00:02<00:10, 1.35s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000249 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657845 valid's binary_logloss: 0.663325
[200] train's binary_logloss: 0.647265 valid's binary_logloss: 0.656205
[300] train's binary_logloss: 0.642195 valid's binary_logloss: 0.653918
[400] train's binary_logloss: 0.638463 valid's binary_logloss: 0.653807
[500] train's binary_logloss: 0.635325 valid's binary_logloss: 0.653223
[600] train's binary_logloss: 0.632547 valid's binary_logloss: 0.652408
[700] train's binary_logloss: 0.629962 valid's binary_logloss: 0.652142
[800] train's binary_logloss: 0.627395 valid's binary_logloss: 0.652339
Early stopping, best iteration is:
[743] train's binary_logloss: 0.628893 valid's binary_logloss: 0.651858
bagging, val_score: 0.651255: 30%|### | 3/10 [00:04<00:09, 1.32s/it][I 2020-09-27 04:39:34,223] Trial 29 finished with value: 0.6518582592610941 and parameters: {'bagging_fraction': 0.8423665601800546, 'bagging_freq': 2}. Best is trial 28 with value: 0.6512546217026939.
bagging, val_score: 0.651255: 30%|### | 3/10 [00:04<00:09, 1.32s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000300 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657643 valid's binary_logloss: 0.663425
[200] train's binary_logloss: 0.647094 valid's binary_logloss: 0.65631
[300] train's binary_logloss: 0.642016 valid's binary_logloss: 0.653742
[400] train's binary_logloss: 0.638407 valid's binary_logloss: 0.653279
Early stopping, best iteration is:
[377] train's binary_logloss: 0.639168 valid's binary_logloss: 0.653138
bagging, val_score: 0.651255: 40%|#### | 4/10 [00:04<00:06, 1.14s/it][I 2020-09-27 04:39:34,939] Trial 30 finished with value: 0.6531380481905572 and parameters: {'bagging_fraction': 0.8282561902576878, 'bagging_freq': 2}. Best is trial 28 with value: 0.6512546217026939.
bagging, val_score: 0.651255: 40%|#### | 4/10 [00:04<00:06, 1.14s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000244 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.65787 valid's binary_logloss: 0.662913
[200] train's binary_logloss: 0.647202 valid's binary_logloss: 0.656389
[300] train's binary_logloss: 0.642181 valid's binary_logloss: 0.65432
[400] train's binary_logloss: 0.638556 valid's binary_logloss: 0.653816
Early stopping, best iteration is:
[361] train's binary_logloss: 0.639969 valid's binary_logloss: 0.653683
bagging, val_score: 0.651255: 50%|##### | 5/10 [00:06<00:05, 1.14s/it][I 2020-09-27 04:39:36,081] Trial 31 finished with value: 0.6536825951065313 and parameters: {'bagging_fraction': 0.8381544500351495, 'bagging_freq': 2}. Best is trial 28 with value: 0.6512546217026939.
bagging, val_score: 0.651255: 50%|##### | 5/10 [00:06<00:05, 1.14s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000455 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657707 valid's binary_logloss: 0.662425
[200] train's binary_logloss: 0.64726 valid's binary_logloss: 0.655364
[300] train's binary_logloss: 0.64231 valid's binary_logloss: 0.653736
[400] train's binary_logloss: 0.638629 valid's binary_logloss: 0.653251
[500] train's binary_logloss: 0.63557 valid's binary_logloss: 0.652996
[600] train's binary_logloss: 0.632836 valid's binary_logloss: 0.653094
[700] train's binary_logloss: 0.630172 valid's binary_logloss: 0.652558
[800] train's binary_logloss: 0.627493 valid's binary_logloss: 0.652618
Early stopping, best iteration is:
[738] train's binary_logloss: 0.629206 valid's binary_logloss: 0.65221
bagging, val_score: 0.651255: 60%|###### | 6/10 [00:07<00:04, 1.16s/it][I 2020-09-27 04:39:37,277] Trial 32 finished with value: 0.6522103084600892 and parameters: {'bagging_fraction': 0.8240973522704712, 'bagging_freq': 2}. Best is trial 28 with value: 0.6512546217026939.
bagging, val_score: 0.651255: 60%|###### | 6/10 [00:07<00:04, 1.16s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000240 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657759 valid's binary_logloss: 0.6633
[200] train's binary_logloss: 0.647172 valid's binary_logloss: 0.656277
[300] train's binary_logloss: 0.642123 valid's binary_logloss: 0.654052
[400] train's binary_logloss: 0.638549 valid's binary_logloss: 0.653547
[500] train's binary_logloss: 0.635603 valid's binary_logloss: 0.653766
Early stopping, best iteration is:
[419] train's binary_logloss: 0.637971 valid's binary_logloss: 0.653468
bagging, val_score: 0.651255: 70%|####### | 7/10 [00:08<00:03, 1.04s/it][I 2020-09-27 04:39:38,055] Trial 33 finished with value: 0.6534681190347887 and parameters: {'bagging_fraction': 0.8133668384521671, 'bagging_freq': 2}. Best is trial 28 with value: 0.6512546217026939.
bagging, val_score: 0.651255: 70%|####### | 7/10 [00:08<00:03, 1.04s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000309 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.65767 valid's binary_logloss: 0.663092
[200] train's binary_logloss: 0.646954 valid's binary_logloss: 0.65622
[300] train's binary_logloss: 0.641882 valid's binary_logloss: 0.653856
[400] train's binary_logloss: 0.638359 valid's binary_logloss: 0.653238
[500] train's binary_logloss: 0.635264 valid's binary_logloss: 0.652485
[600] train's binary_logloss: 0.632569 valid's binary_logloss: 0.652572
Early stopping, best iteration is:
[500] train's binary_logloss: 0.635264 valid's binary_logloss: 0.652485
bagging, val_score: 0.651255: 80%|######## | 8/10 [00:08<00:01, 1.01it/s][I 2020-09-27 04:39:38,927] Trial 34 finished with value: 0.6524845330402806 and parameters: {'bagging_fraction': 0.833189344448115, 'bagging_freq': 2}. Best is trial 28 with value: 0.6512546217026939.
bagging, val_score: 0.651255: 80%|######## | 8/10 [00:08<00:01, 1.01it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000447 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657357 valid's binary_logloss: 0.663082
[200] train's binary_logloss: 0.646962 valid's binary_logloss: 0.656317
[300] train's binary_logloss: 0.642048 valid's binary_logloss: 0.65432
Early stopping, best iteration is:
[285] train's binary_logloss: 0.642706 valid's binary_logloss: 0.653911
bagging, val_score: 0.651255: 90%|######### | 9/10 [00:09<00:01, 1.02s/it][I 2020-09-27 04:39:40,016] Trial 35 finished with value: 0.653910601498707 and parameters: {'bagging_fraction': 0.6953833757673704, 'bagging_freq': 3}. Best is trial 28 with value: 0.6512546217026939.
bagging, val_score: 0.651255: 90%|######### | 9/10 [00:09<00:01, 1.02s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000404 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.658127 valid's binary_logloss: 0.664339
[200] train's binary_logloss: 0.647507 valid's binary_logloss: 0.656812
[300] train's binary_logloss: 0.642315 valid's binary_logloss: 0.65461
[400] train's binary_logloss: 0.638734 valid's binary_logloss: 0.653365
[500] train's binary_logloss: 0.635615 valid's binary_logloss: 0.653259
Early stopping, best iteration is:
[439] train's binary_logloss: 0.637483 valid's binary_logloss: 0.652984
bagging, val_score: 0.651255: 100%|##########| 10/10 [00:10<00:00, 1.03it/s][I 2020-09-27 04:39:40,868] Trial 36 finished with value: 0.6529843808564952 and parameters: {'bagging_fraction': 0.9750898583135423, 'bagging_freq': 5}. Best is trial 28 with value: 0.6512546217026939.
bagging, val_score: 0.651255: 100%|##########| 10/10 [00:10<00:00, 1.08s/it]
feature_fraction_stage2, val_score: 0.651255: 0%| | 0/3 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000412 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657426 valid's binary_logloss: 0.66294
[200] train's binary_logloss: 0.646763 valid's binary_logloss: 0.65579
[300] train's binary_logloss: 0.641842 valid's binary_logloss: 0.653608
[400] train's binary_logloss: 0.638277 valid's binary_logloss: 0.653005
[500] train's binary_logloss: 0.635065 valid's binary_logloss: 0.652871
Early stopping, best iteration is:
[437] train's binary_logloss: 0.637058 valid's binary_logloss: 0.652763
feature_fraction_stage2, val_score: 0.651255: 33%|###3 | 1/3 [00:00<00:01, 1.20it/s][I 2020-09-27 04:39:41,719] Trial 37 finished with value: 0.6527633350213652 and parameters: {'feature_fraction': 0.41600000000000004}. Best is trial 37 with value: 0.6527633350213652.
feature_fraction_stage2, val_score: 0.651255: 33%|###3 | 1/3 [00:00<00:01, 1.20it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000510 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657376 valid's binary_logloss: 0.66298
[200] train's binary_logloss: 0.646944 valid's binary_logloss: 0.656526
[300] train's binary_logloss: 0.641961 valid's binary_logloss: 0.653998
[400] train's binary_logloss: 0.638308 valid's binary_logloss: 0.653033
[500] train's binary_logloss: 0.634963 valid's binary_logloss: 0.652584
[600] train's binary_logloss: 0.632079 valid's binary_logloss: 0.652716
[700] train's binary_logloss: 0.629142 valid's binary_logloss: 0.652471
[800] train's binary_logloss: 0.62641 valid's binary_logloss: 0.652186
Early stopping, best iteration is:
[758] train's binary_logloss: 0.627519 valid's binary_logloss: 0.651784
feature_fraction_stage2, val_score: 0.651255: 67%|######6 | 2/3 [00:02<00:00, 1.00it/s][I 2020-09-27 04:39:43,108] Trial 38 finished with value: 0.6517839548740617 and parameters: {'feature_fraction': 0.44800000000000006}. Best is trial 38 with value: 0.6517839548740617.
feature_fraction_stage2, val_score: 0.651255: 67%|######6 | 2/3 [00:02<00:00, 1.00it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011259 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657376 valid's binary_logloss: 0.66298
[200] train's binary_logloss: 0.646944 valid's binary_logloss: 0.656526
[300] train's binary_logloss: 0.641961 valid's binary_logloss: 0.653998
[400] train's binary_logloss: 0.638308 valid's binary_logloss: 0.653033
[500] train's binary_logloss: 0.634963 valid's binary_logloss: 0.652584
[600] train's binary_logloss: 0.632079 valid's binary_logloss: 0.652716
[700] train's binary_logloss: 0.629142 valid's binary_logloss: 0.652471
[800] train's binary_logloss: 0.62641 valid's binary_logloss: 0.652186
Early stopping, best iteration is:
[758] train's binary_logloss: 0.627519 valid's binary_logloss: 0.651784
feature_fraction_stage2, val_score: 0.651255: 100%|##########| 3/3 [00:03<00:00, 1.15s/it][I 2020-09-27 04:39:44,601] Trial 39 finished with value: 0.6517839548740617 and parameters: {'feature_fraction': 0.48000000000000004}. Best is trial 38 with value: 0.6517839548740617.
feature_fraction_stage2, val_score: 0.651255: 100%|##########| 3/3 [00:03<00:00, 1.24s/it]
regularization_factors, val_score: 0.651255: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000460 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657637 valid's binary_logloss: 0.663101
[200] train's binary_logloss: 0.647154 valid's binary_logloss: 0.656096
[300] train's binary_logloss: 0.642211 valid's binary_logloss: 0.653702
[400] train's binary_logloss: 0.638597 valid's binary_logloss: 0.652862
[500] train's binary_logloss: 0.635489 valid's binary_logloss: 0.652354
[600] train's binary_logloss: 0.632512 valid's binary_logloss: 0.652123
[700] train's binary_logloss: 0.62981 valid's binary_logloss: 0.651852
[800] train's binary_logloss: 0.627227 valid's binary_logloss: 0.651697
Early stopping, best iteration is:
[743] train's binary_logloss: 0.628658 valid's binary_logloss: 0.651255
regularization_factors, val_score: 0.651255: 5%|5 | 1/20 [00:01<00:23, 1.23s/it][I 2020-09-27 04:39:45,848] Trial 40 finished with value: 0.6512546032948433 and parameters: {'lambda_l1': 3.7407304516474708e-06, 'lambda_l2': 0.0003236321792160619}. Best is trial 40 with value: 0.6512546032948433.
regularization_factors, val_score: 0.651255: 5%|5 | 1/20 [00:01<00:23, 1.23s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000479 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657637 valid's binary_logloss: 0.663101
[200] train's binary_logloss: 0.647154 valid's binary_logloss: 0.656096
[300] train's binary_logloss: 0.642211 valid's binary_logloss: 0.653702
[400] train's binary_logloss: 0.638597 valid's binary_logloss: 0.652862
[500] train's binary_logloss: 0.635489 valid's binary_logloss: 0.652354
[600] train's binary_logloss: 0.632512 valid's binary_logloss: 0.652123
[700] train's binary_logloss: 0.62981 valid's binary_logloss: 0.651852
[800] train's binary_logloss: 0.627227 valid's binary_logloss: 0.651697
Early stopping, best iteration is:
[743] train's binary_logloss: 0.628658 valid's binary_logloss: 0.651255
regularization_factors, val_score: 0.651255: 10%|# | 2/20 [00:02<00:23, 1.33s/it][I 2020-09-27 04:39:47,419] Trial 41 finished with value: 0.6512546004999574 and parameters: {'lambda_l1': 1.58423580909119e-06, 'lambda_l2': 0.00037329518153918976}. Best is trial 41 with value: 0.6512546004999574.
regularization_factors, val_score: 0.651255: 10%|# | 2/20 [00:02<00:23, 1.33s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000390 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657637 valid's binary_logloss: 0.663101
[200] train's binary_logloss: 0.647154 valid's binary_logloss: 0.656096
[300] train's binary_logloss: 0.642211 valid's binary_logloss: 0.653702
[400] train's binary_logloss: 0.638597 valid's binary_logloss: 0.652862
[500] train's binary_logloss: 0.635489 valid's binary_logloss: 0.652354
[600] train's binary_logloss: 0.632512 valid's binary_logloss: 0.652123
[700] train's binary_logloss: 0.62981 valid's binary_logloss: 0.651852
[800] train's binary_logloss: 0.627227 valid's binary_logloss: 0.651697
Early stopping, best iteration is:
[743] train's binary_logloss: 0.628658 valid's binary_logloss: 0.651255
regularization_factors, val_score: 0.651255: 15%|#5 | 3/20 [00:04<00:22, 1.33s/it][I 2020-09-27 04:39:48,734] Trial 42 finished with value: 0.6512545924218361 and parameters: {'lambda_l1': 1.3626757884196743e-06, 'lambda_l2': 0.0005160151192105561}. Best is trial 42 with value: 0.6512545924218361.
regularization_factors, val_score: 0.651255: 15%|#5 | 3/20 [00:04<00:22, 1.33s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000381 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657637 valid's binary_logloss: 0.663101
[200] train's binary_logloss: 0.647154 valid's binary_logloss: 0.656096
[300] train's binary_logloss: 0.642211 valid's binary_logloss: 0.653702
[400] train's binary_logloss: 0.638597 valid's binary_logloss: 0.652862
[500] train's binary_logloss: 0.635489 valid's binary_logloss: 0.652354
[600] train's binary_logloss: 0.632512 valid's binary_logloss: 0.652123
[700] train's binary_logloss: 0.62981 valid's binary_logloss: 0.651852
[800] train's binary_logloss: 0.627227 valid's binary_logloss: 0.651697
Early stopping, best iteration is:
[743] train's binary_logloss: 0.628658 valid's binary_logloss: 0.651255
regularization_factors, val_score: 0.651255: 20%|## | 4/20 [00:05<00:20, 1.29s/it][I 2020-09-27 04:39:49,939] Trial 43 finished with value: 0.6512546024195541 and parameters: {'lambda_l1': 1.826384761461434e-06, 'lambda_l2': 0.000339404582931327}. Best is trial 42 with value: 0.6512545924218361.
regularization_factors, val_score: 0.651255: 20%|## | 4/20 [00:05<00:20, 1.29s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000442 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657637 valid's binary_logloss: 0.663101
[200] train's binary_logloss: 0.647154 valid's binary_logloss: 0.656096
[300] train's binary_logloss: 0.642211 valid's binary_logloss: 0.653702
[400] train's binary_logloss: 0.638597 valid's binary_logloss: 0.652862
[500] train's binary_logloss: 0.635489 valid's binary_logloss: 0.652354
[600] train's binary_logloss: 0.632512 valid's binary_logloss: 0.652123
[700] train's binary_logloss: 0.62981 valid's binary_logloss: 0.651852
[800] train's binary_logloss: 0.627227 valid's binary_logloss: 0.651697
Early stopping, best iteration is:
[743] train's binary_logloss: 0.628658 valid's binary_logloss: 0.651255
regularization_factors, val_score: 0.651255: 25%|##5 | 5/20 [00:07<00:21, 1.42s/it][I 2020-09-27 04:39:51,672] Trial 44 finished with value: 0.6512545971181739 and parameters: {'lambda_l1': 9.830713052753886e-07, 'lambda_l2': 0.00043290604666443164}. Best is trial 42 with value: 0.6512545924218361.
regularization_factors, val_score: 0.651255: 25%|##5 | 5/20 [00:07<00:21, 1.42s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000564 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657637 valid's binary_logloss: 0.663101
[200] train's binary_logloss: 0.647154 valid's binary_logloss: 0.656096
[300] train's binary_logloss: 0.642211 valid's binary_logloss: 0.653702
[400] train's binary_logloss: 0.638597 valid's binary_logloss: 0.652862
[500] train's binary_logloss: 0.635489 valid's binary_logloss: 0.652354
[600] train's binary_logloss: 0.632512 valid's binary_logloss: 0.652123
[700] train's binary_logloss: 0.62981 valid's binary_logloss: 0.651852
[800] train's binary_logloss: 0.627227 valid's binary_logloss: 0.651697
Early stopping, best iteration is:
[743] train's binary_logloss: 0.628658 valid's binary_logloss: 0.651255
regularization_factors, val_score: 0.651255: 30%|### | 6/20 [00:08<00:19, 1.36s/it][I 2020-09-27 04:39:52,891] Trial 45 finished with value: 0.6512546006093322 and parameters: {'lambda_l1': 7.770048321005081e-07, 'lambda_l2': 0.00037155770796455103}. Best is trial 42 with value: 0.6512545924218361.
regularization_factors, val_score: 0.651255: 30%|### | 6/20 [00:08<00:19, 1.36s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000386 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657637 valid's binary_logloss: 0.663101
[200] train's binary_logloss: 0.647154 valid's binary_logloss: 0.656096
[300] train's binary_logloss: 0.642211 valid's binary_logloss: 0.653702
[400] train's binary_logloss: 0.638597 valid's binary_logloss: 0.652862
[500] train's binary_logloss: 0.635489 valid's binary_logloss: 0.652354
[600] train's binary_logloss: 0.632512 valid's binary_logloss: 0.652123
[700] train's binary_logloss: 0.62981 valid's binary_logloss: 0.651852
[800] train's binary_logloss: 0.627227 valid's binary_logloss: 0.651697
Early stopping, best iteration is:
[743] train's binary_logloss: 0.628658 valid's binary_logloss: 0.651255
regularization_factors, val_score: 0.651255: 35%|###5 | 7/20 [00:09<00:17, 1.32s/it][I 2020-09-27 04:39:54,130] Trial 46 finished with value: 0.65125458769479 and parameters: {'lambda_l1': 5.65299654936304e-07, 'lambda_l2': 0.0005992895040580498}. Best is trial 46 with value: 0.65125458769479.
regularization_factors, val_score: 0.651255: 35%|###5 | 7/20 [00:09<00:17, 1.32s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000244 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657637 valid's binary_logloss: 0.663101
[200] train's binary_logloss: 0.647154 valid's binary_logloss: 0.656096
[300] train's binary_logloss: 0.642211 valid's binary_logloss: 0.653702
[400] train's binary_logloss: 0.638597 valid's binary_logloss: 0.652862
[500] train's binary_logloss: 0.635489 valid's binary_logloss: 0.652354
[600] train's binary_logloss: 0.632512 valid's binary_logloss: 0.652123
[700] train's binary_logloss: 0.62981 valid's binary_logloss: 0.651852
[800] train's binary_logloss: 0.627227 valid's binary_logloss: 0.651697
Early stopping, best iteration is:
[743] train's binary_logloss: 0.628658 valid's binary_logloss: 0.651255
regularization_factors, val_score: 0.651255: 40%|#### | 8/20 [00:11<00:17, 1.42s/it][I 2020-09-27 04:39:55,765] Trial 47 finished with value: 0.651254582843981 and parameters: {'lambda_l1': 5.113147117976975e-07, 'lambda_l2': 0.0006849316707389379}. Best is trial 47 with value: 0.651254582843981.
regularization_factors, val_score: 0.651255: 40%|#### | 8/20 [00:11<00:17, 1.42s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000452 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657637 valid's binary_logloss: 0.663101
[200] train's binary_logloss: 0.647154 valid's binary_logloss: 0.656096
[300] train's binary_logloss: 0.642211 valid's binary_logloss: 0.653702
[400] train's binary_logloss: 0.638597 valid's binary_logloss: 0.652862
[500] train's binary_logloss: 0.635489 valid's binary_logloss: 0.652354
[600] train's binary_logloss: 0.632512 valid's binary_logloss: 0.652123
[700] train's binary_logloss: 0.62981 valid's binary_logloss: 0.651852
[800] train's binary_logloss: 0.627227 valid's binary_logloss: 0.651697
Early stopping, best iteration is:
[743] train's binary_logloss: 0.628658 valid's binary_logloss: 0.651255
regularization_factors, val_score: 0.651255: 45%|####5 | 9/20 [00:12<00:14, 1.35s/it][I 2020-09-27 04:39:56,969] Trial 48 finished with value: 0.6512545741011545 and parameters: {'lambda_l1': 3.656814382369128e-07, 'lambda_l2': 0.0008392662361736358}. Best is trial 48 with value: 0.6512545741011545.
regularization_factors, val_score: 0.651255: 45%|####5 | 9/20 [00:12<00:14, 1.35s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000266 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657637 valid's binary_logloss: 0.663101
[200] train's binary_logloss: 0.647154 valid's binary_logloss: 0.656096
[300] train's binary_logloss: 0.642211 valid's binary_logloss: 0.653702
[400] train's binary_logloss: 0.638597 valid's binary_logloss: 0.652862
[500] train's binary_logloss: 0.635489 valid's binary_logloss: 0.652354
[600] train's binary_logloss: 0.632513 valid's binary_logloss: 0.652123
[700] train's binary_logloss: 0.629811 valid's binary_logloss: 0.651852
[800] train's binary_logloss: 0.627228 valid's binary_logloss: 0.651696
Early stopping, best iteration is:
[743] train's binary_logloss: 0.628659 valid's binary_logloss: 0.651254
regularization_factors, val_score: 0.651254: 50%|##### | 10/20 [00:13<00:13, 1.31s/it][I 2020-09-27 04:39:58,184] Trial 49 finished with value: 0.65125447186593 and parameters: {'lambda_l1': 4.097168681166982e-08, 'lambda_l2': 0.002642450368328453}. Best is trial 49 with value: 0.65125447186593.
regularization_factors, val_score: 0.651254: 50%|##### | 10/20 [00:13<00:13, 1.31s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000234 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657642 valid's binary_logloss: 0.663104
[200] train's binary_logloss: 0.647161 valid's binary_logloss: 0.6561
[300] train's binary_logloss: 0.642211 valid's binary_logloss: 0.653789
[400] train's binary_logloss: 0.638567 valid's binary_logloss: 0.652995
Early stopping, best iteration is:
[378] train's binary_logloss: 0.639273 valid's binary_logloss: 0.65285
regularization_factors, val_score: 0.651254: 55%|#####5 | 11/20 [00:14<00:11, 1.26s/it][I 2020-09-27 04:39:59,310] Trial 50 finished with value: 0.6528496531447342 and parameters: {'lambda_l1': 1.8926948433638843e-08, 'lambda_l2': 0.18218516904146864}. Best is trial 49 with value: 0.65125447186593.
regularization_factors, val_score: 0.651254: 55%|#####5 | 11/20 [00:14<00:11, 1.26s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000499 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657637 valid's binary_logloss: 0.663101
[200] train's binary_logloss: 0.647154 valid's binary_logloss: 0.656096
[300] train's binary_logloss: 0.642211 valid's binary_logloss: 0.653702
[400] train's binary_logloss: 0.638597 valid's binary_logloss: 0.652862
[500] train's binary_logloss: 0.635489 valid's binary_logloss: 0.652354
[600] train's binary_logloss: 0.632513 valid's binary_logloss: 0.652123
[700] train's binary_logloss: 0.629811 valid's binary_logloss: 0.651852
[800] train's binary_logloss: 0.627228 valid's binary_logloss: 0.651697
Early stopping, best iteration is:
[743] train's binary_logloss: 0.628659 valid's binary_logloss: 0.651254
regularization_factors, val_score: 0.651254: 60%|###### | 12/20 [00:15<00:10, 1.25s/it][I 2020-09-27 04:40:00,559] Trial 51 finished with value: 0.651254491688544 and parameters: {'lambda_l1': 9.464855403688449e-08, 'lambda_l2': 0.0022922152533675836}. Best is trial 49 with value: 0.65125447186593.
regularization_factors, val_score: 0.651254: 60%|###### | 12/20 [00:15<00:10, 1.25s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000381 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657638 valid's binary_logloss: 0.663101
[200] train's binary_logloss: 0.647154 valid's binary_logloss: 0.656096
[300] train's binary_logloss: 0.642213 valid's binary_logloss: 0.653702
[400] train's binary_logloss: 0.6386 valid's binary_logloss: 0.652862
[500] train's binary_logloss: 0.635493 valid's binary_logloss: 0.652354
[600] train's binary_logloss: 0.632518 valid's binary_logloss: 0.652123
[700] train's binary_logloss: 0.629817 valid's binary_logloss: 0.651851
[800] train's binary_logloss: 0.627236 valid's binary_logloss: 0.651695
Early stopping, best iteration is:
[743] train's binary_logloss: 0.628666 valid's binary_logloss: 0.651254
regularization_factors, val_score: 0.651254: 65%|######5 | 13/20 [00:17<00:09, 1.34s/it][I 2020-09-27 04:40:02,095] Trial 52 finished with value: 0.651253816244402 and parameters: {'lambda_l1': 1.1047734679593184e-08, 'lambda_l2': 0.014255881410398455}. Best is trial 52 with value: 0.651253816244402.
regularization_factors, val_score: 0.651254: 65%|######5 | 13/20 [00:17<00:09, 1.34s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000248 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657638 valid's binary_logloss: 0.663102
[200] train's binary_logloss: 0.647156 valid's binary_logloss: 0.656097
[300] train's binary_logloss: 0.642216 valid's binary_logloss: 0.653703
[400] train's binary_logloss: 0.638607 valid's binary_logloss: 0.652863
[500] train's binary_logloss: 0.6355 valid's binary_logloss: 0.652355
[600] train's binary_logloss: 0.632662 valid's binary_logloss: 0.652694
Early stopping, best iteration is:
[531] train's binary_logloss: 0.634567 valid's binary_logloss: 0.652228
regularization_factors, val_score: 0.651254: 70%|####### | 14/20 [00:19<00:09, 1.53s/it][I 2020-09-27 04:40:04,065] Trial 53 finished with value: 0.6522275662505942 and parameters: {'lambda_l1': 1.4618448056501589e-08, 'lambda_l2': 0.04375470564403219}. Best is trial 52 with value: 0.651253816244402.
regularization_factors, val_score: 0.651254: 70%|####### | 14/20 [00:19<00:09, 1.53s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000394 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657638 valid's binary_logloss: 0.663101
[200] train's binary_logloss: 0.647154 valid's binary_logloss: 0.656096
[300] train's binary_logloss: 0.642213 valid's binary_logloss: 0.653702
[400] train's binary_logloss: 0.6386 valid's binary_logloss: 0.652862
[500] train's binary_logloss: 0.635493 valid's binary_logloss: 0.652354
[600] train's binary_logloss: 0.632518 valid's binary_logloss: 0.652123
[700] train's binary_logloss: 0.629817 valid's binary_logloss: 0.651851
[800] train's binary_logloss: 0.627236 valid's binary_logloss: 0.651695
Early stopping, best iteration is:
[743] train's binary_logloss: 0.628666 valid's binary_logloss: 0.651254
regularization_factors, val_score: 0.651254: 75%|#######5 | 15/20 [00:20<00:07, 1.45s/it][I 2020-09-27 04:40:05,343] Trial 54 finished with value: 0.6512538204787435 and parameters: {'lambda_l1': 1.0600410018512844e-08, 'lambda_l2': 0.014180637901064958}. Best is trial 52 with value: 0.651253816244402.
regularization_factors, val_score: 0.651254: 75%|#######5 | 15/20 [00:20<00:07, 1.45s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000400 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657638 valid's binary_logloss: 0.663101
[200] train's binary_logloss: 0.647155 valid's binary_logloss: 0.656096
[300] train's binary_logloss: 0.642214 valid's binary_logloss: 0.653702
[400] train's binary_logloss: 0.638602 valid's binary_logloss: 0.652863
[500] train's binary_logloss: 0.635496 valid's binary_logloss: 0.652353
[600] train's binary_logloss: 0.632522 valid's binary_logloss: 0.652122
[700] train's binary_logloss: 0.629822 valid's binary_logloss: 0.65185
[800] train's binary_logloss: 0.627242 valid's binary_logloss: 0.651695
Early stopping, best iteration is:
[743] train's binary_logloss: 0.628671 valid's binary_logloss: 0.651253
regularization_factors, val_score: 0.651253: 80%|######## | 16/20 [00:22<00:06, 1.51s/it][I 2020-09-27 04:40:06,971] Trial 55 finished with value: 0.6512532800647542 and parameters: {'lambda_l1': 1.1653169345025846e-08, 'lambda_l2': 0.02381661498148718}. Best is trial 55 with value: 0.6512532800647542.
regularization_factors, val_score: 0.651253: 80%|######## | 16/20 [00:22<00:06, 1.51s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.014470 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657638 valid's binary_logloss: 0.663101
[200] train's binary_logloss: 0.647155 valid's binary_logloss: 0.656096
[300] train's binary_logloss: 0.642213 valid's binary_logloss: 0.653702
[400] train's binary_logloss: 0.638601 valid's binary_logloss: 0.652863
[500] train's binary_logloss: 0.635495 valid's binary_logloss: 0.652353
[600] train's binary_logloss: 0.632521 valid's binary_logloss: 0.652122
[700] train's binary_logloss: 0.62982 valid's binary_logloss: 0.65185
[800] train's binary_logloss: 0.62724 valid's binary_logloss: 0.651695
Early stopping, best iteration is:
[743] train's binary_logloss: 0.628669 valid's binary_logloss: 0.651253
regularization_factors, val_score: 0.651253: 85%|########5 | 17/20 [00:23<00:04, 1.46s/it][I 2020-09-27 04:40:08,339] Trial 56 finished with value: 0.6512534743967614 and parameters: {'lambda_l1': 1.0004229968657813e-08, 'lambda_l2': 0.020345235153092046}. Best is trial 55 with value: 0.6512532800647542.
regularization_factors, val_score: 0.651253: 85%|########5 | 17/20 [00:23<00:04, 1.46s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000378 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657638 valid's binary_logloss: 0.663101
[200] train's binary_logloss: 0.647155 valid's binary_logloss: 0.656096
[300] train's binary_logloss: 0.642213 valid's binary_logloss: 0.653702
[400] train's binary_logloss: 0.638602 valid's binary_logloss: 0.652863
[500] train's binary_logloss: 0.635495 valid's binary_logloss: 0.652353
[600] train's binary_logloss: 0.632521 valid's binary_logloss: 0.652122
[700] train's binary_logloss: 0.629821 valid's binary_logloss: 0.65185
[800] train's binary_logloss: 0.62724 valid's binary_logloss: 0.651695
Early stopping, best iteration is:
[743] train's binary_logloss: 0.62867 valid's binary_logloss: 0.651253
regularization_factors, val_score: 0.651253: 90%|######### | 18/20 [00:25<00:02, 1.42s/it][I 2020-09-27 04:40:09,647] Trial 57 finished with value: 0.6512534186191336 and parameters: {'lambda_l1': 1.573306324418443e-08, 'lambda_l2': 0.021340714973965415}. Best is trial 55 with value: 0.6512532800647542.
regularization_factors, val_score: 0.651253: 90%|######### | 18/20 [00:25<00:02, 1.42s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004407 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657638 valid's binary_logloss: 0.663102
[200] train's binary_logloss: 0.647156 valid's binary_logloss: 0.656097
[300] train's binary_logloss: 0.642216 valid's binary_logloss: 0.653703
[400] train's binary_logloss: 0.638606 valid's binary_logloss: 0.652863
[500] train's binary_logloss: 0.635499 valid's binary_logloss: 0.652355
[600] train's binary_logloss: 0.632662 valid's binary_logloss: 0.652694
Early stopping, best iteration is:
[531] train's binary_logloss: 0.634567 valid's binary_logloss: 0.652228
regularization_factors, val_score: 0.651253: 95%|#########5| 19/20 [00:26<00:01, 1.41s/it][I 2020-09-27 04:40:11,033] Trial 58 finished with value: 0.6522275964597894 and parameters: {'lambda_l1': 1.1442661122184645e-08, 'lambda_l2': 0.04237276613246146}. Best is trial 55 with value: 0.6512532800647542.
regularization_factors, val_score: 0.651253: 95%|#########5| 19/20 [00:26<00:01, 1.41s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002336 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657713 valid's binary_logloss: 0.663154
[200] train's binary_logloss: 0.647209 valid's binary_logloss: 0.655905
[300] train's binary_logloss: 0.64221 valid's binary_logloss: 0.65381
[400] train's binary_logloss: 0.638641 valid's binary_logloss: 0.653241
Early stopping, best iteration is:
[351] train's binary_logloss: 0.640312 valid's binary_logloss: 0.65316
regularization_factors, val_score: 0.651253: 100%|##########| 20/20 [00:27<00:00, 1.24s/it][I 2020-09-27 04:40:11,871] Trial 59 finished with value: 0.6531604651549241 and parameters: {'lambda_l1': 0.7583136516531738, 'lambda_l2': 0.017657519269218453}. Best is trial 55 with value: 0.6512532800647542.
regularization_factors, val_score: 0.651253: 100%|##########| 20/20 [00:27<00:00, 1.36s/it]
min_data_in_leaf, val_score: 0.651253: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000390 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657638 valid's binary_logloss: 0.663101
[200] train's binary_logloss: 0.647129 valid's binary_logloss: 0.656026
[300] train's binary_logloss: 0.642165 valid's binary_logloss: 0.653612
[400] train's binary_logloss: 0.638561 valid's binary_logloss: 0.653007
[500] train's binary_logloss: 0.635388 valid's binary_logloss: 0.652537
[600] train's binary_logloss: 0.63262 valid's binary_logloss: 0.652704
Early stopping, best iteration is:
[575] train's binary_logloss: 0.633278 valid's binary_logloss: 0.652405
min_data_in_leaf, val_score: 0.651253: 20%|## | 1/5 [00:01<00:04, 1.04s/it][I 2020-09-27 04:40:12,924] Trial 60 finished with value: 0.6524054109680009 and parameters: {'min_child_samples': 5}. Best is trial 60 with value: 0.6524054109680009.
min_data_in_leaf, val_score: 0.651253: 20%|## | 1/5 [00:01<00:04, 1.04s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000685 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657638 valid's binary_logloss: 0.663101
[200] train's binary_logloss: 0.647173 valid's binary_logloss: 0.656292
[300] train's binary_logloss: 0.64235 valid's binary_logloss: 0.653435
[400] train's binary_logloss: 0.639047 valid's binary_logloss: 0.652707
[500] train's binary_logloss: 0.636172 valid's binary_logloss: 0.651812
[600] train's binary_logloss: 0.633542 valid's binary_logloss: 0.651717
[700] train's binary_logloss: 0.631145 valid's binary_logloss: 0.651603
[800] train's binary_logloss: 0.628853 valid's binary_logloss: 0.651423
Early stopping, best iteration is:
[737] train's binary_logloss: 0.630317 valid's binary_logloss: 0.651187
min_data_in_leaf, val_score: 0.651187: 40%|#### | 2/5 [00:02<00:03, 1.09s/it][I 2020-09-27 04:40:14,129] Trial 61 finished with value: 0.6511874541615081 and parameters: {'min_child_samples': 100}. Best is trial 61 with value: 0.6511874541615081.
min_data_in_leaf, val_score: 0.651187: 40%|#### | 2/5 [00:02<00:03, 1.09s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.006001 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657638 valid's binary_logloss: 0.663101
[200] train's binary_logloss: 0.647155 valid's binary_logloss: 0.656096
[300] train's binary_logloss: 0.64222 valid's binary_logloss: 0.653679
[400] train's binary_logloss: 0.638577 valid's binary_logloss: 0.653154
[500] train's binary_logloss: 0.635489 valid's binary_logloss: 0.652771
[600] train's binary_logloss: 0.632657 valid's binary_logloss: 0.652758
Early stopping, best iteration is:
[531] train's binary_logloss: 0.634621 valid's binary_logloss: 0.65239
min_data_in_leaf, val_score: 0.651187: 60%|###### | 3/5 [00:03<00:02, 1.18s/it][I 2020-09-27 04:40:15,534] Trial 62 finished with value: 0.6523903635256941 and parameters: {'min_child_samples': 25}. Best is trial 61 with value: 0.6511874541615081.
min_data_in_leaf, val_score: 0.651187: 60%|###### | 3/5 [00:03<00:02, 1.18s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000824 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657638 valid's binary_logloss: 0.663101
[200] train's binary_logloss: 0.647129 valid's binary_logloss: 0.656026
[300] train's binary_logloss: 0.642173 valid's binary_logloss: 0.653464
[400] train's binary_logloss: 0.638672 valid's binary_logloss: 0.652989
[500] train's binary_logloss: 0.635515 valid's binary_logloss: 0.652344
[600] train's binary_logloss: 0.632548 valid's binary_logloss: 0.652081
[700] train's binary_logloss: 0.629711 valid's binary_logloss: 0.651529
[800] train's binary_logloss: 0.627105 valid's binary_logloss: 0.651486
Early stopping, best iteration is:
[749] train's binary_logloss: 0.628438 valid's binary_logloss: 0.651112
min_data_in_leaf, val_score: 0.651112: 80%|######## | 4/5 [00:04<00:01, 1.19s/it][I 2020-09-27 04:40:16,745] Trial 63 finished with value: 0.6511117328857268 and parameters: {'min_child_samples': 10}. Best is trial 63 with value: 0.6511117328857268.
min_data_in_leaf, val_score: 0.651112: 80%|######## | 4/5 [00:04<00:01, 1.19s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000400 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4237
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310
[LightGBM] [Info] Start training from score 0.023310
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657638 valid's binary_logloss: 0.663101
[200] train's binary_logloss: 0.647155 valid's binary_logloss: 0.656096
[300] train's binary_logloss: 0.642357 valid's binary_logloss: 0.653658
[400] train's binary_logloss: 0.638938 valid's binary_logloss: 0.653056
[500] train's binary_logloss: 0.636126 valid's binary_logloss: 0.652589
[600] train's binary_logloss: 0.633479 valid's binary_logloss: 0.652767
[700] train's binary_logloss: 0.630836 valid's binary_logloss: 0.652316
Early stopping, best iteration is:
[693] train's binary_logloss: 0.631059 valid's binary_logloss: 0.652118
min_data_in_leaf, val_score: 0.651112: 100%|##########| 5/5 [00:06<00:00, 1.19s/it][I 2020-09-27 04:40:17,925] Trial 64 finished with value: 0.6521175158063565 and parameters: {'min_child_samples': 50}. Best is trial 63 with value: 0.6511117328857268.
min_data_in_leaf, val_score: 0.651112: 100%|##########| 5/5 [00:06<00:00, 1.21s/it]
Fold : 2
[I 2020-09-27 04:40:18,017] A new study created in memory with name: no-name-75b2cbed-db22-45df-bfc2-635871fa6cfc
feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.007450 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.578542 valid's binary_logloss: 0.656491
Early stopping, best iteration is:
[79] train's binary_logloss: 0.59215 valid's binary_logloss: 0.655948
feature_fraction, val_score: 0.655948: 14%|#4 | 1/7 [00:01<00:06, 1.04s/it][I 2020-09-27 04:40:19,072] Trial 0 finished with value: 0.655947777974299 and parameters: {'feature_fraction': 0.5}. Best is trial 0 with value: 0.655947777974299.
feature_fraction, val_score: 0.655948: 14%|#4 | 1/7 [00:01<00:06, 1.04s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.007324 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.572993 valid's binary_logloss: 0.65828
Early stopping, best iteration is:
[79] train's binary_logloss: 0.587161 valid's binary_logloss: 0.656458
feature_fraction, val_score: 0.655948: 29%|##8 | 2/7 [00:01<00:04, 1.12it/s][I 2020-09-27 04:40:19,615] Trial 1 finished with value: 0.6564576518870222 and parameters: {'feature_fraction': 0.8}. Best is trial 0 with value: 0.655947777974299.
feature_fraction, val_score: 0.655948: 29%|##8 | 2/7 [00:01<00:04, 1.12it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.006440 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.570115 valid's binary_logloss: 0.66021
Early stopping, best iteration is:
[64] train's binary_logloss: 0.595593 valid's binary_logloss: 0.658255
feature_fraction, val_score: 0.655948: 43%|####2 | 3/7 [00:02<00:03, 1.29it/s][I 2020-09-27 04:40:20,131] Trial 2 finished with value: 0.6582553269377266 and parameters: {'feature_fraction': 1.0}. Best is trial 0 with value: 0.655947777974299.
feature_fraction, val_score: 0.655948: 43%|####2 | 3/7 [00:02<00:03, 1.29it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010302 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.575364 valid's binary_logloss: 0.659046
Early stopping, best iteration is:
[68] train's binary_logloss: 0.596133 valid's binary_logloss: 0.65791
feature_fraction, val_score: 0.655948: 57%|#####7 | 4/7 [00:02<00:02, 1.47it/s][I 2020-09-27 04:40:20,586] Trial 3 finished with value: 0.6579096104153804 and parameters: {'feature_fraction': 0.7}. Best is trial 0 with value: 0.655947777974299.
feature_fraction, val_score: 0.655948: 57%|#####7 | 4/7 [00:02<00:02, 1.47it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000900 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.572299 valid's binary_logloss: 0.658723
Early stopping, best iteration is:
[74] train's binary_logloss: 0.590022 valid's binary_logloss: 0.656515
feature_fraction, val_score: 0.655948: 71%|#######1 | 5/7 [00:03<00:01, 1.58it/s][I 2020-09-27 04:40:21,113] Trial 4 finished with value: 0.6565150373220784 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 0 with value: 0.655947777974299.
feature_fraction, val_score: 0.655948: 71%|#######1 | 5/7 [00:03<00:01, 1.58it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009831 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.576537 valid's binary_logloss: 0.658655
Early stopping, best iteration is:
[67] train's binary_logloss: 0.598042 valid's binary_logloss: 0.657232
feature_fraction, val_score: 0.655948: 86%|########5 | 6/7 [00:03<00:00, 1.74it/s][I 2020-09-27 04:40:21,552] Trial 5 finished with value: 0.6572319404658664 and parameters: {'feature_fraction': 0.6}. Best is trial 0 with value: 0.655947777974299.
feature_fraction, val_score: 0.655948: 86%|########5 | 6/7 [00:03<00:00, 1.74it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000333 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.583115 valid's binary_logloss: 0.656735
Early stopping, best iteration is:
[85] train's binary_logloss: 0.592606 valid's binary_logloss: 0.655167
feature_fraction, val_score: 0.655167: 100%|##########| 7/7 [00:04<00:00, 1.37it/s][I 2020-09-27 04:40:22,641] Trial 6 finished with value: 0.6551672639343036 and parameters: {'feature_fraction': 0.4}. Best is trial 6 with value: 0.6551672639343036.
feature_fraction, val_score: 0.655167: 100%|##########| 7/7 [00:04<00:00, 1.52it/s]
num_leaves, val_score: 0.655167: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004364 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.362 valid's binary_logloss: 0.666306
Early stopping, best iteration is:
[62] train's binary_logloss: 0.44475 valid's binary_logloss: 0.66107
num_leaves, val_score: 0.655167: 5%|5 | 1/20 [00:00<00:17, 1.09it/s][I 2020-09-27 04:40:23,577] Trial 7 finished with value: 0.6610695796505417 and parameters: {'num_leaves': 186}. Best is trial 7 with value: 0.6610695796505417.
num_leaves, val_score: 0.655167: 5%|5 | 1/20 [00:00<00:17, 1.09it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000378 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.509613 valid's binary_logloss: 0.655007
Early stopping, best iteration is:
[63] train's binary_logloss: 0.555305 valid's binary_logloss: 0.653724
num_leaves, val_score: 0.653724: 10%|# | 2/20 [00:01<00:14, 1.22it/s][I 2020-09-27 04:40:24,159] Trial 8 finished with value: 0.6537238118183929 and parameters: {'num_leaves': 71}. Best is trial 8 with value: 0.6537238118183929.
num_leaves, val_score: 0.653724: 10%|# | 2/20 [00:01<00:14, 1.22it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000384 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.486851 valid's binary_logloss: 0.659048
Early stopping, best iteration is:
[64] train's binary_logloss: 0.536523 valid's binary_logloss: 0.655529
num_leaves, val_score: 0.653724: 15%|#5 | 3/20 [00:02<00:13, 1.29it/s][I 2020-09-27 04:40:24,823] Trial 9 finished with value: 0.6555293293116815 and parameters: {'num_leaves': 85}. Best is trial 8 with value: 0.6537238118183929.
num_leaves, val_score: 0.653724: 15%|#5 | 3/20 [00:02<00:13, 1.29it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000459 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.668828 valid's binary_logloss: 0.671652
[200] train's binary_logloss: 0.659076 valid's binary_logloss: 0.663493
[300] train's binary_logloss: 0.653776 valid's binary_logloss: 0.659129
[400] train's binary_logloss: 0.65062 valid's binary_logloss: 0.656814
[500] train's binary_logloss: 0.648621 valid's binary_logloss: 0.655373
[600] train's binary_logloss: 0.647288 valid's binary_logloss: 0.654268
[700] train's binary_logloss: 0.646365 valid's binary_logloss: 0.653758
[800] train's binary_logloss: 0.645671 valid's binary_logloss: 0.653488
[900] train's binary_logloss: 0.645111 valid's binary_logloss: 0.653271
[1000] train's binary_logloss: 0.644638 valid's binary_logloss: 0.653148
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.644638 valid's binary_logloss: 0.653148
num_leaves, val_score: 0.653148: 20%|## | 4/20 [00:03<00:15, 1.02it/s][I 2020-09-27 04:40:26,293] Trial 10 finished with value: 0.6531478306778341 and parameters: {'num_leaves': 2}. Best is trial 10 with value: 0.6531478306778341.
num_leaves, val_score: 0.653148: 20%|## | 4/20 [00:03<00:15, 1.02it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001421 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.647264 valid's binary_logloss: 0.65894
[200] train's binary_logloss: 0.635718 valid's binary_logloss: 0.655218
[300] train's binary_logloss: 0.627699 valid's binary_logloss: 0.654772
[400] train's binary_logloss: 0.620605 valid's binary_logloss: 0.65463
Early stopping, best iteration is:
[337] train's binary_logloss: 0.625076 valid's binary_logloss: 0.654329
num_leaves, val_score: 0.653148: 25%|##5 | 5/20 [00:04<00:14, 1.05it/s][I 2020-09-27 04:40:27,188] Trial 11 finished with value: 0.6543286934013544 and parameters: {'num_leaves': 5}. Best is trial 10 with value: 0.6531478306778341.
num_leaves, val_score: 0.653148: 25%|##5 | 5/20 [00:04<00:14, 1.05it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000570 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.631583 valid's binary_logloss: 0.655099
[200] train's binary_logloss: 0.611849 valid's binary_logloss: 0.653955
Early stopping, best iteration is:
[189] train's binary_logloss: 0.613791 valid's binary_logloss: 0.653569
num_leaves, val_score: 0.653148: 30%|### | 6/20 [00:05<00:11, 1.22it/s][I 2020-09-27 04:40:27,696] Trial 12 finished with value: 0.6535687066231268 and parameters: {'num_leaves': 10}. Best is trial 10 with value: 0.6531478306778341.
num_leaves, val_score: 0.653148: 30%|### | 6/20 [00:05<00:11, 1.22it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000425 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.634671 valid's binary_logloss: 0.655145
[200] train's binary_logloss: 0.616535 valid's binary_logloss: 0.655474
Early stopping, best iteration is:
[160] train's binary_logloss: 0.62285 valid's binary_logloss: 0.654646
num_leaves, val_score: 0.653148: 35%|###5 | 7/20 [00:05<00:09, 1.39it/s][I 2020-09-27 04:40:28,179] Trial 13 finished with value: 0.6546457645585857 and parameters: {'num_leaves': 9}. Best is trial 10 with value: 0.6531478306778341.
num_leaves, val_score: 0.653148: 35%|###5 | 7/20 [00:05<00:09, 1.39it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000412 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.668828 valid's binary_logloss: 0.671652
[200] train's binary_logloss: 0.659076 valid's binary_logloss: 0.663493
[300] train's binary_logloss: 0.653776 valid's binary_logloss: 0.659129
[400] train's binary_logloss: 0.65062 valid's binary_logloss: 0.656814
[500] train's binary_logloss: 0.648621 valid's binary_logloss: 0.655373
[600] train's binary_logloss: 0.647288 valid's binary_logloss: 0.654268
[700] train's binary_logloss: 0.646365 valid's binary_logloss: 0.653758
[800] train's binary_logloss: 0.645671 valid's binary_logloss: 0.653488
[900] train's binary_logloss: 0.645111 valid's binary_logloss: 0.653271
[1000] train's binary_logloss: 0.644638 valid's binary_logloss: 0.653148
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.644638 valid's binary_logloss: 0.653148
num_leaves, val_score: 0.653148: 40%|#### | 8/20 [00:06<00:10, 1.12it/s][I 2020-09-27 04:40:29,465] Trial 14 finished with value: 0.6531478306778341 and parameters: {'num_leaves': 2}. Best is trial 10 with value: 0.6531478306778341.
num_leaves, val_score: 0.653148: 40%|#### | 8/20 [00:06<00:10, 1.12it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000471 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.306465 valid's binary_logloss: 0.670298
Early stopping, best iteration is:
[44] train's binary_logloss: 0.45679 valid's binary_logloss: 0.663698
num_leaves, val_score: 0.653148: 45%|####5 | 9/20 [00:08<00:12, 1.15s/it][I 2020-09-27 04:40:31,226] Trial 15 finished with value: 0.663697988152403 and parameters: {'num_leaves': 248}. Best is trial 10 with value: 0.6531478306778341.
num_leaves, val_score: 0.653148: 45%|####5 | 9/20 [00:08<00:12, 1.15s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004276 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.520782 valid's binary_logloss: 0.657921
Early stopping, best iteration is:
[85] train's binary_logloss: 0.536614 valid's binary_logloss: 0.656043
num_leaves, val_score: 0.653148: 50%|##### | 10/20 [00:09<00:09, 1.03it/s][I 2020-09-27 04:40:31,778] Trial 16 finished with value: 0.6560430059842651 and parameters: {'num_leaves': 64}. Best is trial 10 with value: 0.6531478306778341.
num_leaves, val_score: 0.653148: 50%|##### | 10/20 [00:09<00:09, 1.03it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000459 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.414427 valid's binary_logloss: 0.657771
Early stopping, best iteration is:
[82] train's binary_logloss: 0.446068 valid's binary_logloss: 0.655491
num_leaves, val_score: 0.653148: 55%|#####5 | 11/20 [00:10<00:08, 1.03it/s][I 2020-09-27 04:40:32,746] Trial 17 finished with value: 0.6554905655085496 and parameters: {'num_leaves': 138}. Best is trial 10 with value: 0.6531478306778341.
num_leaves, val_score: 0.653148: 55%|#####5 | 11/20 [00:10<00:08, 1.03it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000382 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.57465 valid's binary_logloss: 0.656774
[200] train's binary_logloss: 0.519325 valid's binary_logloss: 0.66119
Early stopping, best iteration is:
[115] train's binary_logloss: 0.565412 valid's binary_logloss: 0.656102
num_leaves, val_score: 0.653148: 60%|###### | 12/20 [00:10<00:06, 1.20it/s][I 2020-09-27 04:40:33,269] Trial 18 finished with value: 0.6561023872985662 and parameters: {'num_leaves': 35}. Best is trial 10 with value: 0.6531478306778341.
num_leaves, val_score: 0.653148: 60%|###### | 12/20 [00:10<00:06, 1.20it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000377 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.431599 valid's binary_logloss: 0.661998
Early stopping, best iteration is:
[51] train's binary_logloss: 0.522437 valid's binary_logloss: 0.656248
num_leaves, val_score: 0.653148: 65%|######5 | 13/20 [00:11<00:06, 1.03it/s][I 2020-09-27 04:40:34,569] Trial 19 finished with value: 0.6562478660337796 and parameters: {'num_leaves': 124}. Best is trial 10 with value: 0.6531478306778341.
num_leaves, val_score: 0.653148: 65%|######5 | 13/20 [00:11<00:06, 1.03it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000466 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.563816 valid's binary_logloss: 0.657169
Early stopping, best iteration is:
[88] train's binary_logloss: 0.572908 valid's binary_logloss: 0.656464
num_leaves, val_score: 0.653148: 70%|####### | 14/20 [00:12<00:05, 1.20it/s][I 2020-09-27 04:40:35,073] Trial 20 finished with value: 0.6564643963450036 and parameters: {'num_leaves': 40}. Best is trial 10 with value: 0.6531478306778341.
num_leaves, val_score: 0.653148: 70%|####### | 14/20 [00:12<00:05, 1.20it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000380 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.668828 valid's binary_logloss: 0.671652
[200] train's binary_logloss: 0.659076 valid's binary_logloss: 0.663493
[300] train's binary_logloss: 0.653776 valid's binary_logloss: 0.659129
[400] train's binary_logloss: 0.65062 valid's binary_logloss: 0.656814
[500] train's binary_logloss: 0.648621 valid's binary_logloss: 0.655373
[600] train's binary_logloss: 0.647288 valid's binary_logloss: 0.654268
[700] train's binary_logloss: 0.646365 valid's binary_logloss: 0.653758
[800] train's binary_logloss: 0.645671 valid's binary_logloss: 0.653488
[900] train's binary_logloss: 0.645111 valid's binary_logloss: 0.653271
[1000] train's binary_logloss: 0.644638 valid's binary_logloss: 0.653148
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.644638 valid's binary_logloss: 0.653148
num_leaves, val_score: 0.653148: 75%|#######5 | 15/20 [00:13<00:04, 1.06it/s][I 2020-09-27 04:40:36,259] Trial 21 finished with value: 0.6531478306778341 and parameters: {'num_leaves': 2}. Best is trial 10 with value: 0.6531478306778341.
num_leaves, val_score: 0.653148: 75%|#######5 | 15/20 [00:13<00:04, 1.06it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010841 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.63738 valid's binary_logloss: 0.656329
[200] train's binary_logloss: 0.621173 valid's binary_logloss: 0.654961
[300] train's binary_logloss: 0.608196 valid's binary_logloss: 0.655376
Early stopping, best iteration is:
[225] train's binary_logloss: 0.617766 valid's binary_logloss: 0.654241
num_leaves, val_score: 0.653148: 80%|######## | 16/20 [00:14<00:03, 1.14it/s][I 2020-09-27 04:40:36,999] Trial 22 finished with value: 0.654241444152379 and parameters: {'num_leaves': 8}. Best is trial 10 with value: 0.6531478306778341.
num_leaves, val_score: 0.653148: 80%|######## | 16/20 [00:14<00:03, 1.14it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000584 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.576413 valid's binary_logloss: 0.656352
Early stopping, best iteration is:
[85] train's binary_logloss: 0.58641 valid's binary_logloss: 0.655563
num_leaves, val_score: 0.653148: 85%|########5 | 17/20 [00:15<00:02, 1.02it/s][I 2020-09-27 04:40:38,198] Trial 23 finished with value: 0.6555629950352285 and parameters: {'num_leaves': 34}. Best is trial 10 with value: 0.6531478306778341.
num_leaves, val_score: 0.653148: 85%|########5 | 17/20 [00:15<00:02, 1.02it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000677 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.647264 valid's binary_logloss: 0.65894
[200] train's binary_logloss: 0.635718 valid's binary_logloss: 0.655218
[300] train's binary_logloss: 0.627699 valid's binary_logloss: 0.654772
[400] train's binary_logloss: 0.620605 valid's binary_logloss: 0.65463
Early stopping, best iteration is:
[337] train's binary_logloss: 0.625076 valid's binary_logloss: 0.654329
num_leaves, val_score: 0.653148: 90%|######### | 18/20 [00:16<00:01, 1.07it/s][I 2020-09-27 04:40:39,032] Trial 24 finished with value: 0.6543286934013544 and parameters: {'num_leaves': 5}. Best is trial 10 with value: 0.6531478306778341.
num_leaves, val_score: 0.653148: 90%|######### | 18/20 [00:16<00:01, 1.07it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000412 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.557014 valid's binary_logloss: 0.658549
Early stopping, best iteration is:
[46] train's binary_logloss: 0.606508 valid's binary_logloss: 0.656484
num_leaves, val_score: 0.653148: 95%|#########5| 19/20 [00:16<00:00, 1.23it/s][I 2020-09-27 04:40:39,562] Trial 25 finished with value: 0.6564836527501194 and parameters: {'num_leaves': 44}. Best is trial 10 with value: 0.6531478306778341.
num_leaves, val_score: 0.653148: 95%|#########5| 19/20 [00:16<00:00, 1.23it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000389 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.455025 valid's binary_logloss: 0.660325
Early stopping, best iteration is:
[39] train's binary_logloss: 0.562754 valid's binary_logloss: 0.655052
num_leaves, val_score: 0.653148: 100%|##########| 20/20 [00:17<00:00, 1.29it/s][I 2020-09-27 04:40:40,252] Trial 26 finished with value: 0.655051684902722 and parameters: {'num_leaves': 107}. Best is trial 10 with value: 0.6531478306778341.
num_leaves, val_score: 0.653148: 100%|##########| 20/20 [00:17<00:00, 1.14it/s]
bagging, val_score: 0.653148: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000406 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.668417 valid's binary_logloss: 0.671074
[200] train's binary_logloss: 0.658108 valid's binary_logloss: 0.662187
[300] train's binary_logloss: 0.652572 valid's binary_logloss: 0.657758
[400] train's binary_logloss: 0.649457 valid's binary_logloss: 0.655824
[500] train's binary_logloss: 0.64748 valid's binary_logloss: 0.654202
[600] train's binary_logloss: 0.646194 valid's binary_logloss: 0.653541
[700] train's binary_logloss: 0.645284 valid's binary_logloss: 0.65328
[800] train's binary_logloss: 0.644551 valid's binary_logloss: 0.653263
[900] train's binary_logloss: 0.64391 valid's binary_logloss: 0.653167
[1000] train's binary_logloss: 0.643382 valid's binary_logloss: 0.653181
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.643382 valid's binary_logloss: 0.653181
bagging, val_score: 0.653148: 10%|# | 1/10 [00:01<00:14, 1.60s/it][I 2020-09-27 04:40:41,874] Trial 27 finished with value: 0.6531811973594188 and parameters: {'bagging_fraction': 0.8306103567609042, 'bagging_freq': 5}. Best is trial 27 with value: 0.6531811973594188.
bagging, val_score: 0.653148: 10%|# | 1/10 [00:01<00:14, 1.60s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000338 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.666533 valid's binary_logloss: 0.669291
[200] train's binary_logloss: 0.655913 valid's binary_logloss: 0.659959
[300] train's binary_logloss: 0.650521 valid's binary_logloss: 0.656005
[400] train's binary_logloss: 0.647915 valid's binary_logloss: 0.654692
[500] train's binary_logloss: 0.646471 valid's binary_logloss: 0.65411
[600] train's binary_logloss: 0.645402 valid's binary_logloss: 0.654374
Early stopping, best iteration is:
[511] train's binary_logloss: 0.646347 valid's binary_logloss: 0.653867
bagging, val_score: 0.653148: 20%|## | 2/10 [00:02<00:11, 1.42s/it][I 2020-09-27 04:40:42,870] Trial 28 finished with value: 0.6538669216016616 and parameters: {'bagging_fraction': 0.4517567998193014, 'bagging_freq': 1}. Best is trial 27 with value: 0.6531811973594188.
bagging, val_score: 0.653148: 20%|## | 2/10 [00:02<00:11, 1.42s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000347 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667071 valid's binary_logloss: 0.669326
[200] train's binary_logloss: 0.656588 valid's binary_logloss: 0.660429
[300] train's binary_logloss: 0.651211 valid's binary_logloss: 0.656184
[400] train's binary_logloss: 0.648478 valid's binary_logloss: 0.653762
[500] train's binary_logloss: 0.646843 valid's binary_logloss: 0.653211
[600] train's binary_logloss: 0.645633 valid's binary_logloss: 0.652424
Early stopping, best iteration is:
[573] train's binary_logloss: 0.645934 valid's binary_logloss: 0.652252
bagging, val_score: 0.652252: 30%|### | 3/10 [00:03<00:08, 1.25s/it][I 2020-09-27 04:40:43,705] Trial 29 finished with value: 0.65225160633841 and parameters: {'bagging_fraction': 0.4737096463396763, 'bagging_freq': 7}. Best is trial 29 with value: 0.65225160633841.
bagging, val_score: 0.652252: 30%|### | 3/10 [00:03<00:08, 1.25s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000502 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.666936 valid's binary_logloss: 0.668846
[200] train's binary_logloss: 0.656163 valid's binary_logloss: 0.659914
[300] train's binary_logloss: 0.650959 valid's binary_logloss: 0.655368
[400] train's binary_logloss: 0.648469 valid's binary_logloss: 0.653546
[500] train's binary_logloss: 0.646883 valid's binary_logloss: 0.653403
Early stopping, best iteration is:
[438] train's binary_logloss: 0.647744 valid's binary_logloss: 0.652963
bagging, val_score: 0.652252: 40%|#### | 4/10 [00:04<00:06, 1.07s/it][I 2020-09-27 04:40:44,379] Trial 30 finished with value: 0.6529627183444107 and parameters: {'bagging_fraction': 0.4264803439292804, 'bagging_freq': 7}. Best is trial 29 with value: 0.65225160633841.
bagging, val_score: 0.652252: 40%|#### | 4/10 [00:04<00:06, 1.07s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000349 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.666824 valid's binary_logloss: 0.668808
[200] train's binary_logloss: 0.656165 valid's binary_logloss: 0.65966
[300] train's binary_logloss: 0.650811 valid's binary_logloss: 0.654901
[400] train's binary_logloss: 0.648589 valid's binary_logloss: 0.652893
[500] train's binary_logloss: 0.647086 valid's binary_logloss: 0.653134
[600] train's binary_logloss: 0.645858 valid's binary_logloss: 0.652547
[700] train's binary_logloss: 0.644904 valid's binary_logloss: 0.65326
Early stopping, best iteration is:
[623] train's binary_logloss: 0.64558 valid's binary_logloss: 0.652267
bagging, val_score: 0.652252: 50%|##### | 5/10 [00:05<00:05, 1.18s/it][I 2020-09-27 04:40:45,808] Trial 31 finished with value: 0.6522673041330311 and parameters: {'bagging_fraction': 0.41942184957156203, 'bagging_freq': 7}. Best is trial 29 with value: 0.65225160633841.
bagging, val_score: 0.652252: 50%|##### | 5/10 [00:05<00:05, 1.18s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000470 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.666715 valid's binary_logloss: 0.669238
[200] train's binary_logloss: 0.656103 valid's binary_logloss: 0.660182
[300] train's binary_logloss: 0.650838 valid's binary_logloss: 0.655485
[400] train's binary_logloss: 0.648497 valid's binary_logloss: 0.653732
[500] train's binary_logloss: 0.646909 valid's binary_logloss: 0.653654
Early stopping, best iteration is:
[470] train's binary_logloss: 0.6473 valid's binary_logloss: 0.652896
bagging, val_score: 0.652252: 60%|###### | 6/10 [00:06<00:04, 1.07s/it][I 2020-09-27 04:40:46,607] Trial 32 finished with value: 0.6528955403221239 and parameters: {'bagging_fraction': 0.41256437409812413, 'bagging_freq': 7}. Best is trial 29 with value: 0.65225160633841.
bagging, val_score: 0.652252: 60%|###### | 6/10 [00:06<00:04, 1.07s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000341 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.666696 valid's binary_logloss: 0.669282
[200] train's binary_logloss: 0.656159 valid's binary_logloss: 0.660236
[300] train's binary_logloss: 0.65083 valid's binary_logloss: 0.655589
[400] train's binary_logloss: 0.648563 valid's binary_logloss: 0.653633
[500] train's binary_logloss: 0.64694 valid's binary_logloss: 0.653341
Early stopping, best iteration is:
[470] train's binary_logloss: 0.647303 valid's binary_logloss: 0.652827
bagging, val_score: 0.652252: 70%|####### | 7/10 [00:06<00:02, 1.07it/s][I 2020-09-27 04:40:47,241] Trial 33 finished with value: 0.6528271831886884 and parameters: {'bagging_fraction': 0.40860834213805497, 'bagging_freq': 7}. Best is trial 29 with value: 0.65225160633841.
bagging, val_score: 0.652252: 70%|####### | 7/10 [00:06<00:02, 1.07it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000446 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.666727 valid's binary_logloss: 0.669333
[200] train's binary_logloss: 0.656066 valid's binary_logloss: 0.660295
[300] train's binary_logloss: 0.650735 valid's binary_logloss: 0.655605
[400] train's binary_logloss: 0.648394 valid's binary_logloss: 0.653499
[500] train's binary_logloss: 0.646916 valid's binary_logloss: 0.653049
Early stopping, best iteration is:
[470] train's binary_logloss: 0.647295 valid's binary_logloss: 0.652461
bagging, val_score: 0.652252: 80%|######## | 8/10 [00:07<00:01, 1.15it/s][I 2020-09-27 04:40:47,944] Trial 34 finished with value: 0.6524608085550598 and parameters: {'bagging_fraction': 0.40094624308690235, 'bagging_freq': 7}. Best is trial 29 with value: 0.65225160633841.
bagging, val_score: 0.652252: 80%|######## | 8/10 [00:07<00:01, 1.15it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000704 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.666735 valid's binary_logloss: 0.669024
[200] train's binary_logloss: 0.656104 valid's binary_logloss: 0.660308
[300] train's binary_logloss: 0.650686 valid's binary_logloss: 0.655603
[400] train's binary_logloss: 0.648444 valid's binary_logloss: 0.653985
[500] train's binary_logloss: 0.646852 valid's binary_logloss: 0.653461
Early stopping, best iteration is:
[470] train's binary_logloss: 0.647262 valid's binary_logloss: 0.652979
bagging, val_score: 0.652252: 90%|######### | 9/10 [00:08<00:00, 1.21it/s][I 2020-09-27 04:40:48,677] Trial 35 finished with value: 0.6529787611145468 and parameters: {'bagging_fraction': 0.40896873900368547, 'bagging_freq': 7}. Best is trial 29 with value: 0.65225160633841.
bagging, val_score: 0.652252: 90%|######### | 9/10 [00:08<00:00, 1.21it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010501 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667557 valid's binary_logloss: 0.670365
[200] train's binary_logloss: 0.656945 valid's binary_logloss: 0.660765
[300] train's binary_logloss: 0.651474 valid's binary_logloss: 0.656544
[400] train's binary_logloss: 0.648605 valid's binary_logloss: 0.653988
[500] train's binary_logloss: 0.646848 valid's binary_logloss: 0.653526
[600] train's binary_logloss: 0.645737 valid's binary_logloss: 0.653035
Early stopping, best iteration is:
[588] train's binary_logloss: 0.645878 valid's binary_logloss: 0.652666
bagging, val_score: 0.652252: 100%|##########| 10/10 [00:09<00:00, 1.03it/s][I 2020-09-27 04:40:49,991] Trial 36 finished with value: 0.65266565454767 and parameters: {'bagging_fraction': 0.5936827208546511, 'bagging_freq': 7}. Best is trial 29 with value: 0.65225160633841.
bagging, val_score: 0.652252: 100%|##########| 10/10 [00:09<00:00, 1.03it/s]
feature_fraction_stage2, val_score: 0.652252: 0%| | 0/3 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.004126 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667157 valid's binary_logloss: 0.669629
[200] train's binary_logloss: 0.65666 valid's binary_logloss: 0.660691
[300] train's binary_logloss: 0.651252 valid's binary_logloss: 0.656139
[400] train's binary_logloss: 0.648542 valid's binary_logloss: 0.653738
[500] train's binary_logloss: 0.646904 valid's binary_logloss: 0.653287
[600] train's binary_logloss: 0.645658 valid's binary_logloss: 0.652441
[700] train's binary_logloss: 0.644752 valid's binary_logloss: 0.652982
Early stopping, best iteration is:
[623] train's binary_logloss: 0.645415 valid's binary_logloss: 0.652211
feature_fraction_stage2, val_score: 0.652211: 33%|###3 | 1/3 [00:00<00:01, 1.14it/s][I 2020-09-27 04:40:50,883] Trial 37 finished with value: 0.6522110872949437 and parameters: {'feature_fraction': 0.41600000000000004}. Best is trial 37 with value: 0.6522110872949437.
feature_fraction_stage2, val_score: 0.652211: 33%|###3 | 1/3 [00:00<00:01, 1.14it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000546 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667214 valid's binary_logloss: 0.669614
[200] train's binary_logloss: 0.656661 valid's binary_logloss: 0.660306
[300] train's binary_logloss: 0.651258 valid's binary_logloss: 0.656152
[400] train's binary_logloss: 0.648532 valid's binary_logloss: 0.653868
[500] train's binary_logloss: 0.64686 valid's binary_logloss: 0.65353
[600] train's binary_logloss: 0.645694 valid's binary_logloss: 0.652608
[700] train's binary_logloss: 0.644833 valid's binary_logloss: 0.653223
Early stopping, best iteration is:
[622] train's binary_logloss: 0.645442 valid's binary_logloss: 0.652368
feature_fraction_stage2, val_score: 0.652211: 67%|######6 | 2/3 [00:01<00:00, 1.14it/s][I 2020-09-27 04:40:51,765] Trial 38 finished with value: 0.6523684425922491 and parameters: {'feature_fraction': 0.44800000000000006}. Best is trial 37 with value: 0.6522110872949437.
feature_fraction_stage2, val_score: 0.652211: 67%|######6 | 2/3 [00:01<00:00, 1.14it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000441 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667214 valid's binary_logloss: 0.669614
[200] train's binary_logloss: 0.656661 valid's binary_logloss: 0.660306
[300] train's binary_logloss: 0.651258 valid's binary_logloss: 0.656152
[400] train's binary_logloss: 0.648532 valid's binary_logloss: 0.653868
[500] train's binary_logloss: 0.64686 valid's binary_logloss: 0.65353
[600] train's binary_logloss: 0.645694 valid's binary_logloss: 0.652608
[700] train's binary_logloss: 0.644833 valid's binary_logloss: 0.653223
Early stopping, best iteration is:
[622] train's binary_logloss: 0.645442 valid's binary_logloss: 0.652368
feature_fraction_stage2, val_score: 0.652211: 100%|##########| 3/3 [00:02<00:00, 1.13it/s][I 2020-09-27 04:40:52,662] Trial 39 finished with value: 0.6523684425922491 and parameters: {'feature_fraction': 0.48000000000000004}. Best is trial 37 with value: 0.6522110872949437.
feature_fraction_stage2, val_score: 0.652211: 100%|##########| 3/3 [00:02<00:00, 1.12it/s]
regularization_factors, val_score: 0.652211: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000446 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667157 valid's binary_logloss: 0.669629
[200] train's binary_logloss: 0.65666 valid's binary_logloss: 0.660691
[300] train's binary_logloss: 0.651252 valid's binary_logloss: 0.656139
[400] train's binary_logloss: 0.648542 valid's binary_logloss: 0.653738
[500] train's binary_logloss: 0.646905 valid's binary_logloss: 0.653287
[600] train's binary_logloss: 0.645659 valid's binary_logloss: 0.652441
[700] train's binary_logloss: 0.644752 valid's binary_logloss: 0.652981
Early stopping, best iteration is:
[623] train's binary_logloss: 0.645415 valid's binary_logloss: 0.652211
regularization_factors, val_score: 0.652211: 5%|5 | 1/20 [00:01<00:25, 1.34s/it][I 2020-09-27 04:40:54,031] Trial 40 finished with value: 0.6522109977626771 and parameters: {'lambda_l1': 0.00014164147515972837, 'lambda_l2': 0.005481479902021038}. Best is trial 40 with value: 0.6522109977626771.
regularization_factors, val_score: 0.652211: 5%|5 | 1/20 [00:01<00:25, 1.34s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004749 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667157 valid's binary_logloss: 0.669629
[200] train's binary_logloss: 0.656661 valid's binary_logloss: 0.660691
[300] train's binary_logloss: 0.651253 valid's binary_logloss: 0.656139
[400] train's binary_logloss: 0.648543 valid's binary_logloss: 0.653738
[500] train's binary_logloss: 0.646906 valid's binary_logloss: 0.653287
[600] train's binary_logloss: 0.64566 valid's binary_logloss: 0.652441
[700] train's binary_logloss: 0.644753 valid's binary_logloss: 0.652981
Early stopping, best iteration is:
[623] train's binary_logloss: 0.645416 valid's binary_logloss: 0.652211
regularization_factors, val_score: 0.652211: 10%|# | 2/20 [00:02<00:21, 1.20s/it][I 2020-09-27 04:40:54,895] Trial 41 finished with value: 0.6522109708484218 and parameters: {'lambda_l1': 0.00010302212398099651, 'lambda_l2': 0.007995173172021716}. Best is trial 41 with value: 0.6522109708484218.
regularization_factors, val_score: 0.652211: 10%|# | 2/20 [00:02<00:21, 1.20s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000406 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667157 valid's binary_logloss: 0.669629
[200] train's binary_logloss: 0.656661 valid's binary_logloss: 0.660691
[300] train's binary_logloss: 0.651252 valid's binary_logloss: 0.656139
[400] train's binary_logloss: 0.648542 valid's binary_logloss: 0.653738
[500] train's binary_logloss: 0.646905 valid's binary_logloss: 0.653287
[600] train's binary_logloss: 0.645659 valid's binary_logloss: 0.652441
[700] train's binary_logloss: 0.644752 valid's binary_logloss: 0.652981
Early stopping, best iteration is:
[623] train's binary_logloss: 0.645415 valid's binary_logloss: 0.652211
regularization_factors, val_score: 0.652211: 15%|#5 | 3/20 [00:03<00:18, 1.10s/it][I 2020-09-27 04:40:55,766] Trial 42 finished with value: 0.6522109575004142 and parameters: {'lambda_l1': 5.703427719344543e-05, 'lambda_l2': 0.01133484756475216}. Best is trial 42 with value: 0.6522109575004142.
regularization_factors, val_score: 0.652211: 15%|#5 | 3/20 [00:03<00:18, 1.10s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001530 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667157 valid's binary_logloss: 0.669629
[200] train's binary_logloss: 0.656661 valid's binary_logloss: 0.660691
[300] train's binary_logloss: 0.651252 valid's binary_logloss: 0.656139
[400] train's binary_logloss: 0.648542 valid's binary_logloss: 0.653738
[500] train's binary_logloss: 0.646905 valid's binary_logloss: 0.653287
[600] train's binary_logloss: 0.645659 valid's binary_logloss: 0.652441
[700] train's binary_logloss: 0.644753 valid's binary_logloss: 0.652981
Early stopping, best iteration is:
[623] train's binary_logloss: 0.645416 valid's binary_logloss: 0.652211
regularization_factors, val_score: 0.652211: 20%|## | 4/20 [00:04<00:17, 1.08s/it][I 2020-09-27 04:40:56,793] Trial 43 finished with value: 0.6522109461124184 and parameters: {'lambda_l1': 7.037388835726531e-05, 'lambda_l2': 0.009023034240515858}. Best is trial 43 with value: 0.6522109461124184.
regularization_factors, val_score: 0.652211: 20%|## | 4/20 [00:04<00:17, 1.08s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.017232 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667157 valid's binary_logloss: 0.669629
[200] train's binary_logloss: 0.656661 valid's binary_logloss: 0.660691
[300] train's binary_logloss: 0.651252 valid's binary_logloss: 0.656139
[400] train's binary_logloss: 0.648542 valid's binary_logloss: 0.653738
[500] train's binary_logloss: 0.646905 valid's binary_logloss: 0.653287
[600] train's binary_logloss: 0.645659 valid's binary_logloss: 0.652441
[700] train's binary_logloss: 0.644753 valid's binary_logloss: 0.652981
Early stopping, best iteration is:
[623] train's binary_logloss: 0.645416 valid's binary_logloss: 0.652211
regularization_factors, val_score: 0.652211: 25%|##5 | 5/20 [00:05<00:16, 1.12s/it][I 2020-09-27 04:40:57,991] Trial 44 finished with value: 0.6522108862289743 and parameters: {'lambda_l1': 4.695264971426557e-05, 'lambda_l2': 0.012990997644376091}. Best is trial 44 with value: 0.6522108862289743.
regularization_factors, val_score: 0.652211: 25%|##5 | 5/20 [00:05<00:16, 1.12s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000418 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667157 valid's binary_logloss: 0.669629
[200] train's binary_logloss: 0.656661 valid's binary_logloss: 0.660691
[300] train's binary_logloss: 0.651252 valid's binary_logloss: 0.656139
[400] train's binary_logloss: 0.648542 valid's binary_logloss: 0.653738
[500] train's binary_logloss: 0.646905 valid's binary_logloss: 0.653287
[600] train's binary_logloss: 0.645659 valid's binary_logloss: 0.652441
[700] train's binary_logloss: 0.644753 valid's binary_logloss: 0.652981
Early stopping, best iteration is:
[623] train's binary_logloss: 0.645416 valid's binary_logloss: 0.652211
regularization_factors, val_score: 0.652211: 30%|### | 6/20 [00:06<00:14, 1.05s/it][I 2020-09-27 04:40:58,893] Trial 45 finished with value: 0.6522109405266442 and parameters: {'lambda_l1': 6.628550474869429e-05, 'lambda_l2': 0.00939838589902522}. Best is trial 44 with value: 0.6522108862289743.
regularization_factors, val_score: 0.652211: 30%|### | 6/20 [00:06<00:14, 1.05s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000752 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667157 valid's binary_logloss: 0.669629
[200] train's binary_logloss: 0.656661 valid's binary_logloss: 0.660691
[300] train's binary_logloss: 0.651252 valid's binary_logloss: 0.656139
[400] train's binary_logloss: 0.648542 valid's binary_logloss: 0.653738
[500] train's binary_logloss: 0.646905 valid's binary_logloss: 0.653287
[600] train's binary_logloss: 0.645659 valid's binary_logloss: 0.652441
[700] train's binary_logloss: 0.644753 valid's binary_logloss: 0.652981
Early stopping, best iteration is:
[623] train's binary_logloss: 0.645416 valid's binary_logloss: 0.652211
regularization_factors, val_score: 0.652211: 35%|###5 | 7/20 [00:07<00:13, 1.01s/it][I 2020-09-27 04:40:59,810] Trial 46 finished with value: 0.6522109149315195 and parameters: {'lambda_l1': 5.01427220376058e-05, 'lambda_l2': 0.011108962252046103}. Best is trial 44 with value: 0.6522108862289743.
regularization_factors, val_score: 0.652211: 35%|###5 | 7/20 [00:07<00:13, 1.01s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000545 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667157 valid's binary_logloss: 0.66963
[200] train's binary_logloss: 0.656661 valid's binary_logloss: 0.660691
[300] train's binary_logloss: 0.651253 valid's binary_logloss: 0.656139
[400] train's binary_logloss: 0.648542 valid's binary_logloss: 0.653739
[500] train's binary_logloss: 0.646905 valid's binary_logloss: 0.653287
[600] train's binary_logloss: 0.64566 valid's binary_logloss: 0.652441
[700] train's binary_logloss: 0.644753 valid's binary_logloss: 0.652981
Early stopping, best iteration is:
[623] train's binary_logloss: 0.645416 valid's binary_logloss: 0.652211
regularization_factors, val_score: 0.652211: 40%|#### | 8/20 [00:08<00:12, 1.03s/it][I 2020-09-27 04:41:00,882] Trial 47 finished with value: 0.6522108308126086 and parameters: {'lambda_l1': 4.076885797493761e-05, 'lambda_l2': 0.016633645127292034}. Best is trial 47 with value: 0.6522108308126086.
regularization_factors, val_score: 0.652211: 40%|#### | 8/20 [00:08<00:12, 1.03s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000476 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667157 valid's binary_logloss: 0.66963
[200] train's binary_logloss: 0.656661 valid's binary_logloss: 0.660691
[300] train's binary_logloss: 0.651253 valid's binary_logloss: 0.656139
[400] train's binary_logloss: 0.648543 valid's binary_logloss: 0.653739
[500] train's binary_logloss: 0.646906 valid's binary_logloss: 0.653287
[600] train's binary_logloss: 0.64566 valid's binary_logloss: 0.652441
[700] train's binary_logloss: 0.644754 valid's binary_logloss: 0.652981
Early stopping, best iteration is:
[623] train's binary_logloss: 0.645417 valid's binary_logloss: 0.652211
regularization_factors, val_score: 0.652211: 45%|####5 | 9/20 [00:09<00:11, 1.04s/it][I 2020-09-27 04:41:01,942] Trial 48 finished with value: 0.6522107665018237 and parameters: {'lambda_l1': 4.1633476557442786e-05, 'lambda_l2': 0.020110745479315996}. Best is trial 48 with value: 0.6522107665018237.
regularization_factors, val_score: 0.652211: 45%|####5 | 9/20 [00:09<00:11, 1.04s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000514 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667174 valid's binary_logloss: 0.669606
[200] train's binary_logloss: 0.656682 valid's binary_logloss: 0.660677
[300] train's binary_logloss: 0.651269 valid's binary_logloss: 0.656126
[400] train's binary_logloss: 0.64857 valid's binary_logloss: 0.653596
[500] train's binary_logloss: 0.646941 valid's binary_logloss: 0.653105
[600] train's binary_logloss: 0.645643 valid's binary_logloss: 0.652442
[700] train's binary_logloss: 0.644768 valid's binary_logloss: 0.653208
Early stopping, best iteration is:
[623] train's binary_logloss: 0.645403 valid's binary_logloss: 0.652216
regularization_factors, val_score: 0.652211: 50%|##### | 10/20 [00:10<00:09, 1.02it/s][I 2020-09-27 04:41:02,794] Trial 49 finished with value: 0.6522161063945356 and parameters: {'lambda_l1': 1.0048108252149604e-05, 'lambda_l2': 0.04595411901882225}. Best is trial 48 with value: 0.6522107665018237.
regularization_factors, val_score: 0.652211: 50%|##### | 10/20 [00:10<00:09, 1.02it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000250 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667264 valid's binary_logloss: 0.669374
[200] train's binary_logloss: 0.656765 valid's binary_logloss: 0.660326
[300] train's binary_logloss: 0.651459 valid's binary_logloss: 0.656068
[400] train's binary_logloss: 0.648733 valid's binary_logloss: 0.65386
[500] train's binary_logloss: 0.647116 valid's binary_logloss: 0.653305
[600] train's binary_logloss: 0.645917 valid's binary_logloss: 0.652237
[700] train's binary_logloss: 0.645032 valid's binary_logloss: 0.652691
Early stopping, best iteration is:
[622] train's binary_logloss: 0.645713 valid's binary_logloss: 0.652093
regularization_factors, val_score: 0.652093: 55%|#####5 | 11/20 [00:10<00:08, 1.06it/s][I 2020-09-27 04:41:03,645] Trial 50 finished with value: 0.6520932820946773 and parameters: {'lambda_l1': 5.634084468530463e-06, 'lambda_l2': 2.653747794681544}. Best is trial 50 with value: 0.6520932820946773.
regularization_factors, val_score: 0.652093: 55%|#####5 | 11/20 [00:10<00:08, 1.06it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002392 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667312 valid's binary_logloss: 0.66942
[200] train's binary_logloss: 0.656827 valid's binary_logloss: 0.660345
[300] train's binary_logloss: 0.651466 valid's binary_logloss: 0.656086
[400] train's binary_logloss: 0.648842 valid's binary_logloss: 0.653957
[500] train's binary_logloss: 0.647208 valid's binary_logloss: 0.653403
[600] train's binary_logloss: 0.646029 valid's binary_logloss: 0.652507
[700] train's binary_logloss: 0.645184 valid's binary_logloss: 0.653114
Early stopping, best iteration is:
[622] train's binary_logloss: 0.645818 valid's binary_logloss: 0.652395
regularization_factors, val_score: 0.652093: 60%|###### | 12/20 [00:12<00:08, 1.04s/it][I 2020-09-27 04:41:04,925] Trial 51 finished with value: 0.6523945935028854 and parameters: {'lambda_l1': 7.71584097888483e-06, 'lambda_l2': 4.4305490881206495}. Best is trial 50 with value: 0.6520932820946773.
regularization_factors, val_score: 0.652093: 60%|###### | 12/20 [00:12<00:08, 1.04s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013219 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667193 valid's binary_logloss: 0.669624
[200] train's binary_logloss: 0.656716 valid's binary_logloss: 0.660578
[300] train's binary_logloss: 0.651263 valid's binary_logloss: 0.656044
[400] train's binary_logloss: 0.648561 valid's binary_logloss: 0.654042
[500] train's binary_logloss: 0.646891 valid's binary_logloss: 0.653546
[600] train's binary_logloss: 0.645722 valid's binary_logloss: 0.652543
[700] train's binary_logloss: 0.644815 valid's binary_logloss: 0.653225
Early stopping, best iteration is:
[623] train's binary_logloss: 0.645497 valid's binary_logloss: 0.652335
regularization_factors, val_score: 0.652093: 65%|######5 | 13/20 [00:13<00:08, 1.16s/it][I 2020-09-27 04:41:06,349] Trial 52 finished with value: 0.6523351333047752 and parameters: {'lambda_l1': 7.422157693379738e-06, 'lambda_l2': 0.7332370123602728}. Best is trial 50 with value: 0.6520932820946773.
regularization_factors, val_score: 0.652093: 65%|######5 | 13/20 [00:13<00:08, 1.16s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000421 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667345 valid's binary_logloss: 0.66941
[200] train's binary_logloss: 0.656873 valid's binary_logloss: 0.660467
[300] train's binary_logloss: 0.651513 valid's binary_logloss: 0.655838
[400] train's binary_logloss: 0.648818 valid's binary_logloss: 0.653846
[500] train's binary_logloss: 0.647188 valid's binary_logloss: 0.653186
[600] train's binary_logloss: 0.646098 valid's binary_logloss: 0.652539
Early stopping, best iteration is:
[573] train's binary_logloss: 0.646377 valid's binary_logloss: 0.652312
regularization_factors, val_score: 0.652093: 70%|####### | 14/20 [00:14<00:06, 1.07s/it][I 2020-09-27 04:41:07,202] Trial 53 finished with value: 0.6523123313741976 and parameters: {'lambda_l1': 1.6736112176387592, 'lambda_l2': 3.1184950223667215e-07}. Best is trial 50 with value: 0.6520932820946773.
regularization_factors, val_score: 0.652093: 70%|####### | 14/20 [00:14<00:06, 1.07s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002056 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667157 valid's binary_logloss: 0.669629
[200] train's binary_logloss: 0.65666 valid's binary_logloss: 0.660691
[300] train's binary_logloss: 0.651252 valid's binary_logloss: 0.656139
[400] train's binary_logloss: 0.648542 valid's binary_logloss: 0.653738
[500] train's binary_logloss: 0.646905 valid's binary_logloss: 0.653287
[600] train's binary_logloss: 0.645659 valid's binary_logloss: 0.652441
[700] train's binary_logloss: 0.644752 valid's binary_logloss: 0.652981
Early stopping, best iteration is:
[623] train's binary_logloss: 0.645415 valid's binary_logloss: 0.652211
regularization_factors, val_score: 0.652093: 75%|#######5 | 15/20 [00:15<00:05, 1.02s/it][I 2020-09-27 04:41:08,102] Trial 54 finished with value: 0.6522110253635155 and parameters: {'lambda_l1': 0.0016228255112486223, 'lambda_l2': 8.007222095393149e-05}. Best is trial 50 with value: 0.6520932820946773.
regularization_factors, val_score: 0.652093: 75%|#######5 | 15/20 [00:15<00:05, 1.02s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000382 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667157 valid's binary_logloss: 0.669629
[200] train's binary_logloss: 0.65666 valid's binary_logloss: 0.660691
[300] train's binary_logloss: 0.651252 valid's binary_logloss: 0.656139
[400] train's binary_logloss: 0.648542 valid's binary_logloss: 0.653738
[500] train's binary_logloss: 0.646904 valid's binary_logloss: 0.653287
[600] train's binary_logloss: 0.645658 valid's binary_logloss: 0.652441
[700] train's binary_logloss: 0.644752 valid's binary_logloss: 0.652982
Early stopping, best iteration is:
[623] train's binary_logloss: 0.645415 valid's binary_logloss: 0.652211
regularization_factors, val_score: 0.652093: 80%|######## | 16/20 [00:16<00:04, 1.09s/it][I 2020-09-27 04:41:09,376] Trial 55 finished with value: 0.6522110831327725 and parameters: {'lambda_l1': 8.98931163949704e-08, 'lambda_l2': 0.0002694496140730187}. Best is trial 50 with value: 0.6520932820946773.
regularization_factors, val_score: 0.652093: 80%|######## | 16/20 [00:16<00:04, 1.09s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000250 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.66718 valid's binary_logloss: 0.669612
[200] train's binary_logloss: 0.656691 valid's binary_logloss: 0.660683
[300] train's binary_logloss: 0.651278 valid's binary_logloss: 0.656131
[400] train's binary_logloss: 0.648583 valid's binary_logloss: 0.653603
[500] train's binary_logloss: 0.646957 valid's binary_logloss: 0.653104
[600] train's binary_logloss: 0.645691 valid's binary_logloss: 0.652458
[700] train's binary_logloss: 0.644799 valid's binary_logloss: 0.653165
Early stopping, best iteration is:
[623] train's binary_logloss: 0.645445 valid's binary_logloss: 0.65223
regularization_factors, val_score: 0.652093: 85%|########5 | 17/20 [00:17<00:03, 1.03s/it][I 2020-09-27 04:41:10,268] Trial 56 finished with value: 0.6522296062194786 and parameters: {'lambda_l1': 1.272264047860099e-06, 'lambda_l2': 0.29472834124394065}. Best is trial 50 with value: 0.6520932820946773.
regularization_factors, val_score: 0.652093: 85%|########5 | 17/20 [00:17<00:03, 1.03s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000520 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667157 valid's binary_logloss: 0.669629
[200] train's binary_logloss: 0.656661 valid's binary_logloss: 0.660691
[300] train's binary_logloss: 0.651252 valid's binary_logloss: 0.656139
[400] train's binary_logloss: 0.648542 valid's binary_logloss: 0.653738
[500] train's binary_logloss: 0.646905 valid's binary_logloss: 0.653287
[600] train's binary_logloss: 0.645659 valid's binary_logloss: 0.652441
[700] train's binary_logloss: 0.644752 valid's binary_logloss: 0.652981
Early stopping, best iteration is:
[623] train's binary_logloss: 0.645415 valid's binary_logloss: 0.652211
regularization_factors, val_score: 0.652093: 90%|######### | 18/20 [00:18<00:01, 1.02it/s][I 2020-09-27 04:41:11,137] Trial 57 finished with value: 0.6522109880071963 and parameters: {'lambda_l1': 0.0023192016556508074, 'lambda_l2': 0.0008178552198766434}. Best is trial 50 with value: 0.6520932820946773.
regularization_factors, val_score: 0.652093: 90%|######### | 18/20 [00:18<00:01, 1.02it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000580 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667176 valid's binary_logloss: 0.669608
[200] train's binary_logloss: 0.656684 valid's binary_logloss: 0.660679
[300] train's binary_logloss: 0.651272 valid's binary_logloss: 0.656127
[400] train's binary_logloss: 0.648574 valid's binary_logloss: 0.653598
[500] train's binary_logloss: 0.646946 valid's binary_logloss: 0.653105
[600] train's binary_logloss: 0.645649 valid's binary_logloss: 0.652439
[700] train's binary_logloss: 0.644774 valid's binary_logloss: 0.653205
Early stopping, best iteration is:
[623] train's binary_logloss: 0.64541 valid's binary_logloss: 0.652214
regularization_factors, val_score: 0.652093: 95%|#########5| 19/20 [00:19<00:00, 1.05it/s][I 2020-09-27 04:41:12,010] Trial 58 finished with value: 0.6522142117867799 and parameters: {'lambda_l1': 0.0007602407297528348, 'lambda_l2': 0.11500147374550874}. Best is trial 50 with value: 0.6520932820946773.
regularization_factors, val_score: 0.652093: 95%|#########5| 19/20 [00:19<00:00, 1.05it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005008 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667156 valid's binary_logloss: 0.669629
[200] train's binary_logloss: 0.656659 valid's binary_logloss: 0.66069
[300] train's binary_logloss: 0.651251 valid's binary_logloss: 0.656139
[400] train's binary_logloss: 0.64854 valid's binary_logloss: 0.653738
[500] train's binary_logloss: 0.646903 valid's binary_logloss: 0.653287
[600] train's binary_logloss: 0.645657 valid's binary_logloss: 0.652441
[700] train's binary_logloss: 0.64475 valid's binary_logloss: 0.652981
Early stopping, best iteration is:
[623] train's binary_logloss: 0.645413 valid's binary_logloss: 0.652211
regularization_factors, val_score: 0.652093: 100%|##########| 20/20 [00:20<00:00, 1.07s/it][I 2020-09-27 04:41:13,360] Trial 59 finished with value: 0.6522109875781879 and parameters: {'lambda_l1': 1.3713971348787328e-06, 'lambda_l2': 0.0013479533247432005}. Best is trial 50 with value: 0.6520932820946773.
regularization_factors, val_score: 0.652093: 100%|##########| 20/20 [00:20<00:00, 1.03s/it]
min_data_in_leaf, val_score: 0.652093: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004509 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667264 valid's binary_logloss: 0.669374
[200] train's binary_logloss: 0.656765 valid's binary_logloss: 0.660326
[300] train's binary_logloss: 0.651459 valid's binary_logloss: 0.656068
[400] train's binary_logloss: 0.648733 valid's binary_logloss: 0.653788
[500] train's binary_logloss: 0.647111 valid's binary_logloss: 0.653197
[600] train's binary_logloss: 0.645996 valid's binary_logloss: 0.652398
Early stopping, best iteration is:
[589] train's binary_logloss: 0.646096 valid's binary_logloss: 0.652267
min_data_in_leaf, val_score: 0.652093: 20%|## | 1/5 [00:00<00:03, 1.23it/s][I 2020-09-27 04:41:14,186] Trial 60 finished with value: 0.6522671792511908 and parameters: {'min_child_samples': 50}. Best is trial 60 with value: 0.6522671792511908.
min_data_in_leaf, val_score: 0.652093: 20%|## | 1/5 [00:00<00:03, 1.23it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000404 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667264 valid's binary_logloss: 0.669374
[200] train's binary_logloss: 0.656765 valid's binary_logloss: 0.660326
[300] train's binary_logloss: 0.651459 valid's binary_logloss: 0.656068
[400] train's binary_logloss: 0.648733 valid's binary_logloss: 0.653788
[500] train's binary_logloss: 0.647134 valid's binary_logloss: 0.653299
[600] train's binary_logloss: 0.645934 valid's binary_logloss: 0.652389
Early stopping, best iteration is:
[573] train's binary_logloss: 0.646258 valid's binary_logloss: 0.652282
min_data_in_leaf, val_score: 0.652093: 40%|#### | 2/5 [00:01<00:02, 1.24it/s][I 2020-09-27 04:41:14,982] Trial 61 finished with value: 0.6522824409223347 and parameters: {'min_child_samples': 25}. Best is trial 60 with value: 0.6522671792511908.
min_data_in_leaf, val_score: 0.652093: 40%|#### | 2/5 [00:01<00:02, 1.24it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001187 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667264 valid's binary_logloss: 0.669374
[200] train's binary_logloss: 0.656765 valid's binary_logloss: 0.660326
[300] train's binary_logloss: 0.651459 valid's binary_logloss: 0.656068
[400] train's binary_logloss: 0.648733 valid's binary_logloss: 0.65386
[500] train's binary_logloss: 0.647145 valid's binary_logloss: 0.653231
[600] train's binary_logloss: 0.645937 valid's binary_logloss: 0.652184
Early stopping, best iteration is:
[573] train's binary_logloss: 0.646242 valid's binary_logloss: 0.652
min_data_in_leaf, val_score: 0.652000: 60%|###### | 3/5 [00:02<00:01, 1.25it/s][I 2020-09-27 04:41:15,775] Trial 62 finished with value: 0.6520002540523853 and parameters: {'min_child_samples': 10}. Best is trial 62 with value: 0.6520002540523853.
min_data_in_leaf, val_score: 0.652000: 60%|###### | 3/5 [00:02<00:01, 1.25it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000591 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667264 valid's binary_logloss: 0.669374
[200] train's binary_logloss: 0.656765 valid's binary_logloss: 0.660326
[300] train's binary_logloss: 0.651468 valid's binary_logloss: 0.656076
[400] train's binary_logloss: 0.648844 valid's binary_logloss: 0.65385
[500] train's binary_logloss: 0.647295 valid's binary_logloss: 0.653583
[600] train's binary_logloss: 0.646236 valid's binary_logloss: 0.652591
Early stopping, best iteration is:
[572] train's binary_logloss: 0.64651 valid's binary_logloss: 0.652484
min_data_in_leaf, val_score: 0.652000: 80%|######## | 4/5 [00:03<00:00, 1.07it/s][I 2020-09-27 04:41:17,010] Trial 63 finished with value: 0.6524841574858531 and parameters: {'min_child_samples': 100}. Best is trial 62 with value: 0.6520002540523853.
min_data_in_leaf, val_score: 0.652000: 80%|######## | 4/5 [00:03<00:00, 1.07it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000499 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002
[LightGBM] [Info] Start training from score 0.023002
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667264 valid's binary_logloss: 0.669374
[200] train's binary_logloss: 0.656765 valid's binary_logloss: 0.660326
[300] train's binary_logloss: 0.651459 valid's binary_logloss: 0.656068
[400] train's binary_logloss: 0.648733 valid's binary_logloss: 0.65386
[500] train's binary_logloss: 0.647145 valid's binary_logloss: 0.653231
[600] train's binary_logloss: 0.645937 valid's binary_logloss: 0.652184
Early stopping, best iteration is:
[573] train's binary_logloss: 0.646242 valid's binary_logloss: 0.652
min_data_in_leaf, val_score: 0.652000: 100%|##########| 5/5 [00:04<00:00, 1.13it/s][I 2020-09-27 04:41:17,796] Trial 64 finished with value: 0.6520002540523853 and parameters: {'min_child_samples': 5}. Best is trial 62 with value: 0.6520002540523853.
min_data_in_leaf, val_score: 0.652000: 100%|##########| 5/5 [00:04<00:00, 1.13it/s]
Fold : 3
[I 2020-09-27 04:41:17,860] A new study created in memory with name: no-name-2f38a2a7-25fd-480a-a0ca-a8d760f2fdde
feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004894 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.572639 valid's binary_logloss: 0.657549
Early stopping, best iteration is:
[69] train's binary_logloss: 0.593294 valid's binary_logloss: 0.65643
feature_fraction, val_score: 0.656430: 14%|#4 | 1/7 [00:00<00:03, 1.89it/s][I 2020-09-27 04:41:18,402] Trial 0 finished with value: 0.6564300006090903 and parameters: {'feature_fraction': 0.8}. Best is trial 0 with value: 0.6564300006090903.
feature_fraction, val_score: 0.656430: 14%|#4 | 1/7 [00:00<00:03, 1.89it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000471 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.582961 valid's binary_logloss: 0.657331
[200] train's binary_logloss: 0.532881 valid's binary_logloss: 0.663509
Early stopping, best iteration is:
[105] train's binary_logloss: 0.579867 valid's binary_logloss: 0.657057
feature_fraction, val_score: 0.656430: 29%|##8 | 2/7 [00:01<00:02, 1.88it/s][I 2020-09-27 04:41:18,934] Trial 1 finished with value: 0.6570566397890953 and parameters: {'feature_fraction': 0.4}. Best is trial 0 with value: 0.6564300006090903.
feature_fraction, val_score: 0.656430: 29%|##8 | 2/7 [00:01<00:02, 1.88it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000466 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.578663 valid's binary_logloss: 0.658469
Early stopping, best iteration is:
[94] train's binary_logloss: 0.582397 valid's binary_logloss: 0.657964
feature_fraction, val_score: 0.656430: 43%|####2 | 3/7 [00:01<00:02, 1.85it/s][I 2020-09-27 04:41:19,494] Trial 2 finished with value: 0.6579636929665533 and parameters: {'feature_fraction': 0.5}. Best is trial 0 with value: 0.6564300006090903.
feature_fraction, val_score: 0.656430: 43%|####2 | 3/7 [00:01<00:02, 1.85it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.008636 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.571407 valid's binary_logloss: 0.662124
Early stopping, best iteration is:
[73] train's binary_logloss: 0.589006 valid's binary_logloss: 0.660046
feature_fraction, val_score: 0.656430: 57%|#####7 | 4/7 [00:02<00:02, 1.47it/s][I 2020-09-27 04:41:20,502] Trial 3 finished with value: 0.6600459491612519 and parameters: {'feature_fraction': 1.0}. Best is trial 0 with value: 0.6564300006090903.
feature_fraction, val_score: 0.656430: 57%|#####7 | 4/7 [00:02<00:02, 1.47it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.007591 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.576825 valid's binary_logloss: 0.659351
Early stopping, best iteration is:
[59] train's binary_logloss: 0.60446 valid's binary_logloss: 0.656858
feature_fraction, val_score: 0.656430: 71%|#######1 | 5/7 [00:03<00:01, 1.57it/s][I 2020-09-27 04:41:21,037] Trial 4 finished with value: 0.6568575848518106 and parameters: {'feature_fraction': 0.6}. Best is trial 0 with value: 0.6564300006090903.
feature_fraction, val_score: 0.656430: 71%|#######1 | 5/7 [00:03<00:01, 1.57it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.005521 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.574018 valid's binary_logloss: 0.658815
Early stopping, best iteration is:
[58] train's binary_logloss: 0.603435 valid's binary_logloss: 0.65692
feature_fraction, val_score: 0.656430: 86%|########5 | 6/7 [00:03<00:00, 1.70it/s][I 2020-09-27 04:41:21,513] Trial 5 finished with value: 0.6569204528754933 and parameters: {'feature_fraction': 0.7}. Best is trial 0 with value: 0.6564300006090903.
feature_fraction, val_score: 0.656430: 86%|########5 | 6/7 [00:03<00:00, 1.70it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000901 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.571583 valid's binary_logloss: 0.66348
Early stopping, best iteration is:
[58] train's binary_logloss: 0.601331 valid's binary_logloss: 0.660277
feature_fraction, val_score: 0.656430: 100%|##########| 7/7 [00:04<00:00, 1.81it/s][I 2020-09-27 04:41:21,986] Trial 6 finished with value: 0.660277034185044 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 0 with value: 0.6564300006090903.
feature_fraction, val_score: 0.656430: 100%|##########| 7/7 [00:04<00:00, 1.70it/s]
num_leaves, val_score: 0.656430: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000435 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.316367 valid's binary_logloss: 0.678402
Early stopping, best iteration is:
[38] train's binary_logloss: 0.477737 valid's binary_logloss: 0.661617
num_leaves, val_score: 0.656430: 5%|5 | 1/20 [00:01<00:22, 1.18s/it][I 2020-09-27 04:41:23,181] Trial 7 finished with value: 0.6616171717247548 and parameters: {'num_leaves': 199}. Best is trial 7 with value: 0.6616171717247548.
num_leaves, val_score: 0.656430: 5%|5 | 1/20 [00:01<00:22, 1.18s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000849 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.294294 valid's binary_logloss: 0.683498
Early stopping, best iteration is:
[30] train's binary_logloss: 0.496866 valid's binary_logloss: 0.663982
num_leaves, val_score: 0.656430: 10%|# | 2/20 [00:02<00:24, 1.34s/it][I 2020-09-27 04:41:24,901] Trial 8 finished with value: 0.6639821465726022 and parameters: {'num_leaves': 220}. Best is trial 7 with value: 0.6616171717247548.
num_leaves, val_score: 0.656430: 10%|# | 2/20 [00:02<00:24, 1.34s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000432 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.317927 valid's binary_logloss: 0.67117
Early stopping, best iteration is:
[38] train's binary_logloss: 0.478124 valid's binary_logloss: 0.658535
num_leaves, val_score: 0.656430: 15%|#5 | 3/20 [00:04<00:21, 1.28s/it][I 2020-09-27 04:41:26,051] Trial 9 finished with value: 0.6585349270193589 and parameters: {'num_leaves': 197}. Best is trial 9 with value: 0.6585349270193589.
num_leaves, val_score: 0.656430: 15%|#5 | 3/20 [00:04<00:21, 1.28s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000872 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.6214 valid's binary_logloss: 0.657582
[200] train's binary_logloss: 0.595975 valid's binary_logloss: 0.659137
Early stopping, best iteration is:
[112] train's binary_logloss: 0.618201 valid's binary_logloss: 0.656843
num_leaves, val_score: 0.656430: 20%|## | 4/20 [00:04<00:16, 1.04s/it][I 2020-09-27 04:41:26,504] Trial 10 finished with value: 0.6568431632806528 and parameters: {'num_leaves': 12}. Best is trial 10 with value: 0.6568431632806528.
num_leaves, val_score: 0.656430: 20%|## | 4/20 [00:04<00:16, 1.04s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000825 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.633772 valid's binary_logloss: 0.658331
[200] train's binary_logloss: 0.615857 valid's binary_logloss: 0.657932
Early stopping, best iteration is:
[151] train's binary_logloss: 0.623812 valid's binary_logloss: 0.657048
num_leaves, val_score: 0.656430: 25%|##5 | 5/20 [00:04<00:12, 1.16it/s][I 2020-09-27 04:41:26,971] Trial 11 finished with value: 0.6570482164914353 and parameters: {'num_leaves': 8}. Best is trial 10 with value: 0.6568431632806528.
num_leaves, val_score: 0.656430: 25%|##5 | 5/20 [00:04<00:12, 1.16it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008860 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.6214 valid's binary_logloss: 0.657582
[200] train's binary_logloss: 0.595975 valid's binary_logloss: 0.659137
Early stopping, best iteration is:
[112] train's binary_logloss: 0.618201 valid's binary_logloss: 0.656843
num_leaves, val_score: 0.656430: 30%|### | 6/20 [00:05<00:11, 1.18it/s][I 2020-09-27 04:41:27,770] Trial 12 finished with value: 0.6568431632806528 and parameters: {'num_leaves': 12}. Best is trial 10 with value: 0.6568431632806528.
num_leaves, val_score: 0.656430: 30%|### | 6/20 [00:05<00:11, 1.18it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012578 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.458511 valid's binary_logloss: 0.667979
Early stopping, best iteration is:
[42] train's binary_logloss: 0.55365 valid's binary_logloss: 0.659615
num_leaves, val_score: 0.656430: 35%|###5 | 7/20 [00:06<00:11, 1.14it/s][I 2020-09-27 04:41:28,720] Trial 13 finished with value: 0.6596145393331994 and parameters: {'num_leaves': 89}. Best is trial 10 with value: 0.6568431632806528.
num_leaves, val_score: 0.656430: 35%|###5 | 7/20 [00:06<00:11, 1.14it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007578 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.458372 valid's binary_logloss: 0.664586
Early stopping, best iteration is:
[37] train's binary_logloss: 0.562981 valid's binary_logloss: 0.659102
num_leaves, val_score: 0.656430: 40%|#### | 8/20 [00:07<00:09, 1.23it/s][I 2020-09-27 04:41:29,393] Trial 14 finished with value: 0.6591016497023785 and parameters: {'num_leaves': 90}. Best is trial 10 with value: 0.6568431632806528.
num_leaves, val_score: 0.656430: 40%|#### | 8/20 [00:07<00:09, 1.23it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001036 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.500134 valid's binary_logloss: 0.660323
Early stopping, best iteration is:
[55] train's binary_logloss: 0.556151 valid's binary_logloss: 0.658021
num_leaves, val_score: 0.656430: 45%|####5 | 9/20 [00:08<00:08, 1.29it/s][I 2020-09-27 04:41:30,070] Trial 15 finished with value: 0.6580206440647851 and parameters: {'num_leaves': 66}. Best is trial 10 with value: 0.6568431632806528.
num_leaves, val_score: 0.656430: 45%|####5 | 9/20 [00:08<00:08, 1.29it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004995 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.37816 valid's binary_logloss: 0.67245
Early stopping, best iteration is:
[33] train's binary_logloss: 0.531324 valid's binary_logloss: 0.658413
num_leaves, val_score: 0.656430: 50%|##### | 10/20 [00:08<00:07, 1.27it/s][I 2020-09-27 04:41:30,896] Trial 16 finished with value: 0.6584130655795914 and parameters: {'num_leaves': 144}. Best is trial 10 with value: 0.6568431632806528.
num_leaves, val_score: 0.656430: 50%|##### | 10/20 [00:08<00:07, 1.27it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000912 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.557347 valid's binary_logloss: 0.658946
Early stopping, best iteration is:
[52] train's binary_logloss: 0.597034 valid's binary_logloss: 0.657578
num_leaves, val_score: 0.656430: 55%|#####5 | 11/20 [00:09<00:07, 1.15it/s][I 2020-09-27 04:41:31,958] Trial 17 finished with value: 0.6575775290564063 and parameters: {'num_leaves': 38}. Best is trial 10 with value: 0.6568431632806528.
num_leaves, val_score: 0.656430: 55%|#####5 | 11/20 [00:09<00:07, 1.15it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004745 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.380846 valid's binary_logloss: 0.673729
Early stopping, best iteration is:
[29] train's binary_logloss: 0.545732 valid's binary_logloss: 0.661242
num_leaves, val_score: 0.656430: 60%|###### | 12/20 [00:10<00:06, 1.17it/s][I 2020-09-27 04:41:32,785] Trial 18 finished with value: 0.6612424428440397 and parameters: {'num_leaves': 143}. Best is trial 10 with value: 0.6568431632806528.
num_leaves, val_score: 0.656430: 60%|###### | 12/20 [00:10<00:06, 1.17it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000446 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.270716 valid's binary_logloss: 0.684739
Early stopping, best iteration is:
[22] train's binary_logloss: 0.522558 valid's binary_logloss: 0.66216
num_leaves, val_score: 0.656430: 65%|######5 | 13/20 [00:12<00:06, 1.03it/s][I 2020-09-27 04:41:34,008] Trial 19 finished with value: 0.6621600912726282 and parameters: {'num_leaves': 248}. Best is trial 10 with value: 0.6568431632806528.
num_leaves, val_score: 0.656430: 65%|######5 | 13/20 [00:12<00:06, 1.03it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000999 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.53747 valid's binary_logloss: 0.659534
Early stopping, best iteration is:
[53] train's binary_logloss: 0.583686 valid's binary_logloss: 0.65689
num_leaves, val_score: 0.656430: 70%|####### | 14/20 [00:12<00:05, 1.18it/s][I 2020-09-27 04:41:34,577] Trial 20 finished with value: 0.6568899584326288 and parameters: {'num_leaves': 47}. Best is trial 10 with value: 0.6568431632806528.
num_leaves, val_score: 0.656430: 70%|####### | 14/20 [00:12<00:05, 1.18it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000858 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.668556 valid's binary_logloss: 0.673758
[200] train's binary_logloss: 0.658927 valid's binary_logloss: 0.665036
[300] train's binary_logloss: 0.653688 valid's binary_logloss: 0.660827
[400] train's binary_logloss: 0.650562 valid's binary_logloss: 0.658072
[500] train's binary_logloss: 0.648564 valid's binary_logloss: 0.656465
[600] train's binary_logloss: 0.647233 valid's binary_logloss: 0.655679
[700] train's binary_logloss: 0.646297 valid's binary_logloss: 0.655125
[800] train's binary_logloss: 0.64559 valid's binary_logloss: 0.654927
[900] train's binary_logloss: 0.645037 valid's binary_logloss: 0.654748
[1000] train's binary_logloss: 0.644582 valid's binary_logloss: 0.654744
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.644582 valid's binary_logloss: 0.654744
num_leaves, val_score: 0.654744: 75%|#######5 | 15/20 [00:14<00:05, 1.10s/it][I 2020-09-27 04:41:36,250] Trial 21 finished with value: 0.6547443275104798 and parameters: {'num_leaves': 2}. Best is trial 21 with value: 0.6547443275104798.
num_leaves, val_score: 0.654744: 75%|#######5 | 15/20 [00:14<00:05, 1.10s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005030 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.607455 valid's binary_logloss: 0.658232
Early stopping, best iteration is:
[89] train's binary_logloss: 0.611867 valid's binary_logloss: 0.657658
num_leaves, val_score: 0.654744: 80%|######## | 16/20 [00:14<00:03, 1.11it/s][I 2020-09-27 04:41:36,691] Trial 22 finished with value: 0.6576583665143456 and parameters: {'num_leaves': 17}. Best is trial 21 with value: 0.6547443275104798.
num_leaves, val_score: 0.654744: 80%|######## | 16/20 [00:14<00:03, 1.11it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000984 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657455 valid's binary_logloss: 0.665049
[200] train's binary_logloss: 0.647037 valid's binary_logloss: 0.658049
[300] train's binary_logloss: 0.641503 valid's binary_logloss: 0.655542
[400] train's binary_logloss: 0.6376 valid's binary_logloss: 0.6548
[500] train's binary_logloss: 0.634488 valid's binary_logloss: 0.654825
[600] train's binary_logloss: 0.631676 valid's binary_logloss: 0.654591
Early stopping, best iteration is:
[550] train's binary_logloss: 0.633043 valid's binary_logloss: 0.654226
num_leaves, val_score: 0.654226: 85%|########5 | 17/20 [00:15<00:02, 1.11it/s][I 2020-09-27 04:41:37,600] Trial 23 finished with value: 0.6542264931984162 and parameters: {'num_leaves': 3}. Best is trial 23 with value: 0.6542264931984162.
num_leaves, val_score: 0.654226: 85%|########5 | 17/20 [00:15<00:02, 1.11it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000878 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.44017 valid's binary_logloss: 0.669582
Early stopping, best iteration is:
[27] train's binary_logloss: 0.579264 valid's binary_logloss: 0.660699
num_leaves, val_score: 0.654226: 90%|######### | 18/20 [00:16<00:01, 1.17it/s][I 2020-09-27 04:41:38,351] Trial 24 finished with value: 0.6606988142412215 and parameters: {'num_leaves': 101}. Best is trial 23 with value: 0.6542264931984162.
num_leaves, val_score: 0.654226: 90%|######### | 18/20 [00:16<00:01, 1.17it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000476 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.534907 valid's binary_logloss: 0.660247
Early stopping, best iteration is:
[52] train's binary_logloss: 0.583748 valid's binary_logloss: 0.656588
num_leaves, val_score: 0.654226: 95%|#########5| 19/20 [00:17<00:00, 1.05it/s][I 2020-09-27 04:41:39,536] Trial 25 finished with value: 0.6565875567036938 and parameters: {'num_leaves': 48}. Best is trial 23 with value: 0.6542264931984162.
num_leaves, val_score: 0.654226: 95%|#########5| 19/20 [00:17<00:00, 1.05it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000989 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.412133 valid's binary_logloss: 0.670637
Early stopping, best iteration is:
[35] train's binary_logloss: 0.542676 valid's binary_logloss: 0.662727
num_leaves, val_score: 0.654226: 100%|##########| 20/20 [00:18<00:00, 1.09it/s][I 2020-09-27 04:41:40,357] Trial 26 finished with value: 0.6627267041691768 and parameters: {'num_leaves': 120}. Best is trial 23 with value: 0.6542264931984162.
num_leaves, val_score: 0.654226: 100%|##########| 20/20 [00:18<00:00, 1.09it/s]
bagging, val_score: 0.654226: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004771 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656611 valid's binary_logloss: 0.664545
[200] train's binary_logloss: 0.64587 valid's binary_logloss: 0.657646
[300] train's binary_logloss: 0.640645 valid's binary_logloss: 0.655721
[400] train's binary_logloss: 0.636853 valid's binary_logloss: 0.655725
[500] train's binary_logloss: 0.633496 valid's binary_logloss: 0.655542
[600] train's binary_logloss: 0.630016 valid's binary_logloss: 0.655279
Early stopping, best iteration is:
[583] train's binary_logloss: 0.6307 valid's binary_logloss: 0.655229
bagging, val_score: 0.654226: 10%|# | 1/10 [00:00<00:08, 1.05it/s][I 2020-09-27 04:41:41,322] Trial 27 finished with value: 0.6552294862634005 and parameters: {'bagging_fraction': 0.7419038111841914, 'bagging_freq': 4}. Best is trial 27 with value: 0.6552294862634005.
bagging, val_score: 0.654226: 10%|# | 1/10 [00:00<00:08, 1.05it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011398 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656644 valid's binary_logloss: 0.664211
[200] train's binary_logloss: 0.646379 valid's binary_logloss: 0.65746
[300] train's binary_logloss: 0.64107 valid's binary_logloss: 0.655347
[400] train's binary_logloss: 0.63725 valid's binary_logloss: 0.654521
[500] train's binary_logloss: 0.633892 valid's binary_logloss: 0.654747
Early stopping, best iteration is:
[418] train's binary_logloss: 0.636564 valid's binary_logloss: 0.654327
bagging, val_score: 0.654226: 20%|## | 2/10 [00:01<00:07, 1.10it/s][I 2020-09-27 04:41:42,144] Trial 28 finished with value: 0.6543274423941562 and parameters: {'bagging_fraction': 0.7471333362321411, 'bagging_freq': 4}. Best is trial 28 with value: 0.6543274423941562.
bagging, val_score: 0.654226: 20%|## | 2/10 [00:01<00:07, 1.10it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000868 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657402 valid's binary_logloss: 0.66564
[200] train's binary_logloss: 0.646902 valid's binary_logloss: 0.659056
[300] train's binary_logloss: 0.641447 valid's binary_logloss: 0.656325
[400] train's binary_logloss: 0.637573 valid's binary_logloss: 0.655624
[500] train's binary_logloss: 0.634342 valid's binary_logloss: 0.655314
Early stopping, best iteration is:
[451] train's binary_logloss: 0.635853 valid's binary_logloss: 0.655055
bagging, val_score: 0.654226: 30%|### | 3/10 [00:03<00:07, 1.05s/it][I 2020-09-27 04:41:43,507] Trial 29 finished with value: 0.6550545130006391 and parameters: {'bagging_fraction': 0.9853823451672025, 'bagging_freq': 7}. Best is trial 28 with value: 0.6543274423941562.
bagging, val_score: 0.654226: 30%|### | 3/10 [00:03<00:07, 1.05s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005058 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.655834 valid's binary_logloss: 0.664357
[200] train's binary_logloss: 0.645749 valid's binary_logloss: 0.656264
[300] train's binary_logloss: 0.641025 valid's binary_logloss: 0.656599
Early stopping, best iteration is:
[244] train's binary_logloss: 0.643386 valid's binary_logloss: 0.656058
bagging, val_score: 0.654226: 40%|#### | 4/10 [00:03<00:05, 1.08it/s][I 2020-09-27 04:41:44,148] Trial 30 finished with value: 0.6560583873045698 and parameters: {'bagging_fraction': 0.5579781104550179, 'bagging_freq': 3}. Best is trial 28 with value: 0.6543274423941562.
bagging, val_score: 0.654226: 40%|#### | 4/10 [00:03<00:05, 1.08it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000466 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657297 valid's binary_logloss: 0.665521
[200] train's binary_logloss: 0.64689 valid's binary_logloss: 0.658315
[300] train's binary_logloss: 0.641484 valid's binary_logloss: 0.655464
[400] train's binary_logloss: 0.637638 valid's binary_logloss: 0.654412
[500] train's binary_logloss: 0.634281 valid's binary_logloss: 0.653965
[600] train's binary_logloss: 0.631536 valid's binary_logloss: 0.653942
Early stopping, best iteration is:
[511] train's binary_logloss: 0.633982 valid's binary_logloss: 0.65385
bagging, val_score: 0.653850: 50%|##### | 5/10 [00:04<00:04, 1.06it/s][I 2020-09-27 04:41:45,139] Trial 31 finished with value: 0.6538504332512516 and parameters: {'bagging_fraction': 0.9521342276139986, 'bagging_freq': 7}. Best is trial 31 with value: 0.6538504332512516.
bagging, val_score: 0.653850: 50%|##### | 5/10 [00:04<00:04, 1.06it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005270 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657127 valid's binary_logloss: 0.665533
[200] train's binary_logloss: 0.646571 valid's binary_logloss: 0.65777
[300] train's binary_logloss: 0.641217 valid's binary_logloss: 0.655631
[400] train's binary_logloss: 0.637412 valid's binary_logloss: 0.654782
[500] train's binary_logloss: 0.63418 valid's binary_logloss: 0.654189
Early stopping, best iteration is:
[483] train's binary_logloss: 0.634685 valid's binary_logloss: 0.654112
bagging, val_score: 0.653850: 60%|###### | 6/10 [00:05<00:03, 1.07it/s][I 2020-09-27 04:41:46,049] Trial 32 finished with value: 0.6541122278782455 and parameters: {'bagging_fraction': 0.9284471640035712, 'bagging_freq': 7}. Best is trial 31 with value: 0.6538504332512516.
bagging, val_score: 0.653850: 60%|###### | 6/10 [00:05<00:03, 1.07it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000855 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657282 valid's binary_logloss: 0.665245
[200] train's binary_logloss: 0.646764 valid's binary_logloss: 0.657757
[300] train's binary_logloss: 0.641399 valid's binary_logloss: 0.655443
[400] train's binary_logloss: 0.637639 valid's binary_logloss: 0.654744
[500] train's binary_logloss: 0.634399 valid's binary_logloss: 0.654233
[600] train's binary_logloss: 0.63167 valid's binary_logloss: 0.654358
Early stopping, best iteration is:
[506] train's binary_logloss: 0.634248 valid's binary_logloss: 0.654145
bagging, val_score: 0.653850: 70%|####### | 7/10 [00:07<00:03, 1.05s/it][I 2020-09-27 04:41:47,378] Trial 33 finished with value: 0.6541452383331995 and parameters: {'bagging_fraction': 0.9500310046900615, 'bagging_freq': 7}. Best is trial 31 with value: 0.6538504332512516.
bagging, val_score: 0.653850: 70%|####### | 7/10 [00:07<00:03, 1.05s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004962 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657608 valid's binary_logloss: 0.665788
[200] train's binary_logloss: 0.647027 valid's binary_logloss: 0.6581
[300] train's binary_logloss: 0.641573 valid's binary_logloss: 0.655658
[400] train's binary_logloss: 0.637741 valid's binary_logloss: 0.655284
[500] train's binary_logloss: 0.634569 valid's binary_logloss: 0.654728
[600] train's binary_logloss: 0.631817 valid's binary_logloss: 0.654394
[700] train's binary_logloss: 0.629035 valid's binary_logloss: 0.654408
Early stopping, best iteration is:
[653] train's binary_logloss: 0.630298 valid's binary_logloss: 0.654176
bagging, val_score: 0.653850: 80%|######## | 8/10 [00:08<00:02, 1.11s/it][I 2020-09-27 04:41:48,627] Trial 34 finished with value: 0.654175795227107 and parameters: {'bagging_fraction': 0.9963869875397509, 'bagging_freq': 7}. Best is trial 31 with value: 0.6538504332512516.
bagging, val_score: 0.653850: 80%|######## | 8/10 [00:08<00:02, 1.11s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005318 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657371 valid's binary_logloss: 0.665888
[200] train's binary_logloss: 0.646844 valid's binary_logloss: 0.658962
[300] train's binary_logloss: 0.64144 valid's binary_logloss: 0.656351
[400] train's binary_logloss: 0.637614 valid's binary_logloss: 0.654985
[500] train's binary_logloss: 0.634424 valid's binary_logloss: 0.654861
[600] train's binary_logloss: 0.631665 valid's binary_logloss: 0.654356
[700] train's binary_logloss: 0.628979 valid's binary_logloss: 0.654652
Early stopping, best iteration is:
[625] train's binary_logloss: 0.631017 valid's binary_logloss: 0.654177
bagging, val_score: 0.653850: 90%|######### | 9/10 [00:09<00:01, 1.19s/it][I 2020-09-27 04:41:50,003] Trial 35 finished with value: 0.6541765913434219 and parameters: {'bagging_fraction': 0.9914234683174175, 'bagging_freq': 7}. Best is trial 31 with value: 0.6538504332512516.
bagging, val_score: 0.653850: 90%|######### | 9/10 [00:09<00:01, 1.19s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.006921 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657129 valid's binary_logloss: 0.666433
[200] train's binary_logloss: 0.646683 valid's binary_logloss: 0.658903
[300] train's binary_logloss: 0.641282 valid's binary_logloss: 0.656802
[400] train's binary_logloss: 0.637336 valid's binary_logloss: 0.655707
[500] train's binary_logloss: 0.634119 valid's binary_logloss: 0.654705
[600] train's binary_logloss: 0.631418 valid's binary_logloss: 0.654631
Early stopping, best iteration is:
[510] train's binary_logloss: 0.633847 valid's binary_logloss: 0.654396
bagging, val_score: 0.653850: 100%|##########| 10/10 [00:11<00:00, 1.28s/it][I 2020-09-27 04:41:51,487] Trial 36 finished with value: 0.6543958028999971 and parameters: {'bagging_fraction': 0.9429983760903378, 'bagging_freq': 7}. Best is trial 31 with value: 0.6538504332512516.
bagging, val_score: 0.653850: 100%|##########| 10/10 [00:11<00:00, 1.11s/it]
feature_fraction_stage2, val_score: 0.653850: 0%| | 0/6 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000804 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657244 valid's binary_logloss: 0.665676
[200] train's binary_logloss: 0.646834 valid's binary_logloss: 0.657933
[300] train's binary_logloss: 0.641472 valid's binary_logloss: 0.655835
[400] train's binary_logloss: 0.637734 valid's binary_logloss: 0.654847
[500] train's binary_logloss: 0.634612 valid's binary_logloss: 0.654743
[600] train's binary_logloss: 0.631834 valid's binary_logloss: 0.654606
Early stopping, best iteration is:
[559] train's binary_logloss: 0.632955 valid's binary_logloss: 0.654387
feature_fraction_stage2, val_score: 0.653850: 17%|#6 | 1/6 [00:01<00:05, 1.06s/it][I 2020-09-27 04:41:52,568] Trial 37 finished with value: 0.6543873568694546 and parameters: {'feature_fraction': 0.7520000000000001}. Best is trial 37 with value: 0.6543873568694546.
feature_fraction_stage2, val_score: 0.653850: 17%|#6 | 1/6 [00:01<00:05, 1.06s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001058 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657307 valid's binary_logloss: 0.665041
[200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725
[300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506
[400] train's binary_logloss: 0.637619 valid's binary_logloss: 0.654377
[500] train's binary_logloss: 0.63434 valid's binary_logloss: 0.653884
Early stopping, best iteration is:
[488] train's binary_logloss: 0.63474 valid's binary_logloss: 0.653747
feature_fraction_stage2, val_score: 0.653747: 33%|###3 | 2/6 [00:02<00:04, 1.04s/it][I 2020-09-27 04:41:53,565] Trial 38 finished with value: 0.6537470788868942 and parameters: {'feature_fraction': 0.7200000000000001}. Best is trial 38 with value: 0.6537470788868942.
feature_fraction_stage2, val_score: 0.653747: 33%|###3 | 2/6 [00:02<00:04, 1.04s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001682 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657318 valid's binary_logloss: 0.665806
[200] train's binary_logloss: 0.646625 valid's binary_logloss: 0.658644
[300] train's binary_logloss: 0.641194 valid's binary_logloss: 0.656049
[400] train's binary_logloss: 0.63737 valid's binary_logloss: 0.655556
[500] train's binary_logloss: 0.634009 valid's binary_logloss: 0.654948
[600] train's binary_logloss: 0.63129 valid's binary_logloss: 0.65525
Early stopping, best iteration is:
[557] train's binary_logloss: 0.63246 valid's binary_logloss: 0.654813
feature_fraction_stage2, val_score: 0.653747: 50%|##### | 3/6 [00:03<00:03, 1.21s/it][I 2020-09-27 04:41:55,150] Trial 39 finished with value: 0.6548128899258291 and parameters: {'feature_fraction': 0.8480000000000001}. Best is trial 38 with value: 0.6537470788868942.
feature_fraction_stage2, val_score: 0.653747: 50%|##### | 3/6 [00:03<00:03, 1.21s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001085 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657207 valid's binary_logloss: 0.665659
[200] train's binary_logloss: 0.646755 valid's binary_logloss: 0.657975
[300] train's binary_logloss: 0.641365 valid's binary_logloss: 0.655752
[400] train's binary_logloss: 0.637538 valid's binary_logloss: 0.654478
[500] train's binary_logloss: 0.634276 valid's binary_logloss: 0.654404
[600] train's binary_logloss: 0.63152 valid's binary_logloss: 0.654509
Early stopping, best iteration is:
[533] train's binary_logloss: 0.633346 valid's binary_logloss: 0.654136
feature_fraction_stage2, val_score: 0.653747: 67%|######6 | 4/6 [00:04<00:02, 1.22s/it][I 2020-09-27 04:41:56,409] Trial 40 finished with value: 0.6541357158243993 and parameters: {'feature_fraction': 0.88}. Best is trial 38 with value: 0.6537470788868942.
feature_fraction_stage2, val_score: 0.653747: 67%|######6 | 4/6 [00:04<00:02, 1.22s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005676 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657297 valid's binary_logloss: 0.665521
[200] train's binary_logloss: 0.64689 valid's binary_logloss: 0.658315
[300] train's binary_logloss: 0.641484 valid's binary_logloss: 0.655464
[400] train's binary_logloss: 0.637638 valid's binary_logloss: 0.654412
[500] train's binary_logloss: 0.634281 valid's binary_logloss: 0.653965
[600] train's binary_logloss: 0.631536 valid's binary_logloss: 0.653942
Early stopping, best iteration is:
[511] train's binary_logloss: 0.633982 valid's binary_logloss: 0.65385
feature_fraction_stage2, val_score: 0.653747: 83%|########3 | 5/6 [00:05<00:01, 1.18s/it][I 2020-09-27 04:41:57,501] Trial 41 finished with value: 0.6538504332512516 and parameters: {'feature_fraction': 0.8160000000000001}. Best is trial 38 with value: 0.6537470788868942.
feature_fraction_stage2, val_score: 0.653747: 83%|########3 | 5/6 [00:06<00:01, 1.18s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010335 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657244 valid's binary_logloss: 0.665676
[200] train's binary_logloss: 0.646834 valid's binary_logloss: 0.657933
[300] train's binary_logloss: 0.641472 valid's binary_logloss: 0.655835
[400] train's binary_logloss: 0.637734 valid's binary_logloss: 0.654847
[500] train's binary_logloss: 0.634612 valid's binary_logloss: 0.654743
[600] train's binary_logloss: 0.631834 valid's binary_logloss: 0.654606
Early stopping, best iteration is:
[559] train's binary_logloss: 0.632955 valid's binary_logloss: 0.654387
feature_fraction_stage2, val_score: 0.653747: 100%|##########| 6/6 [00:07<00:00, 1.29s/it][I 2020-09-27 04:41:59,040] Trial 42 finished with value: 0.6543873568694546 and parameters: {'feature_fraction': 0.784}. Best is trial 38 with value: 0.6537470788868942.
feature_fraction_stage2, val_score: 0.653747: 100%|##########| 6/6 [00:07<00:00, 1.26s/it]
regularization_factors, val_score: 0.653747: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000984 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657307 valid's binary_logloss: 0.665041
[200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725
[300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506
[400] train's binary_logloss: 0.637619 valid's binary_logloss: 0.654377
[500] train's binary_logloss: 0.63434 valid's binary_logloss: 0.653884
Early stopping, best iteration is:
[488] train's binary_logloss: 0.63474 valid's binary_logloss: 0.653747
regularization_factors, val_score: 0.653747: 5%|5 | 1/20 [00:00<00:17, 1.11it/s][I 2020-09-27 04:41:59,955] Trial 43 finished with value: 0.6537470788623818 and parameters: {'lambda_l1': 1.164724543974479e-08, 'lambda_l2': 6.922005451695355e-06}. Best is trial 43 with value: 0.6537470788623818.
regularization_factors, val_score: 0.653747: 5%|5 | 1/20 [00:00<00:17, 1.11it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000880 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657307 valid's binary_logloss: 0.665041
[200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725
[300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506
[400] train's binary_logloss: 0.637619 valid's binary_logloss: 0.654377
[500] train's binary_logloss: 0.63434 valid's binary_logloss: 0.653884
Early stopping, best iteration is:
[488] train's binary_logloss: 0.63474 valid's binary_logloss: 0.653747
regularization_factors, val_score: 0.653747: 10%|# | 2/20 [00:01<00:16, 1.11it/s][I 2020-09-27 04:42:00,859] Trial 44 finished with value: 0.6537470788688746 and parameters: {'lambda_l1': 1.2101619670742207e-08, 'lambda_l2': 3.569458567905685e-06}. Best is trial 43 with value: 0.6537470788623818.
regularization_factors, val_score: 0.653747: 10%|# | 2/20 [00:01<00:16, 1.11it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000804 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657307 valid's binary_logloss: 0.665041
[200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725
[300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506
[400] train's binary_logloss: 0.637619 valid's binary_logloss: 0.654377
[500] train's binary_logloss: 0.63434 valid's binary_logloss: 0.653884
Early stopping, best iteration is:
[488] train's binary_logloss: 0.63474 valid's binary_logloss: 0.653747
regularization_factors, val_score: 0.653747: 15%|#5 | 3/20 [00:02<00:15, 1.08it/s][I 2020-09-27 04:42:01,825] Trial 45 finished with value: 0.653747078870228 and parameters: {'lambda_l1': 1.0924605897873326e-08, 'lambda_l2': 3.7438094337593657e-06}. Best is trial 43 with value: 0.6537470788623818.
regularization_factors, val_score: 0.653747: 15%|#5 | 3/20 [00:02<00:15, 1.08it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009826 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657307 valid's binary_logloss: 0.665041
[200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725
[300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506
[400] train's binary_logloss: 0.637619 valid's binary_logloss: 0.654377
[500] train's binary_logloss: 0.63434 valid's binary_logloss: 0.653884
Early stopping, best iteration is:
[488] train's binary_logloss: 0.63474 valid's binary_logloss: 0.653747
regularization_factors, val_score: 0.653747: 20%|## | 4/20 [00:04<00:20, 1.28s/it][I 2020-09-27 04:42:03,949] Trial 46 finished with value: 0.6537470788718757 and parameters: {'lambda_l1': 1.479258071574677e-08, 'lambda_l2': 3.0264187607450267e-06}. Best is trial 43 with value: 0.6537470788623818.
regularization_factors, val_score: 0.653747: 20%|## | 4/20 [00:04<00:20, 1.28s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004927 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657307 valid's binary_logloss: 0.665041
[200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725
[300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506
[400] train's binary_logloss: 0.637619 valid's binary_logloss: 0.654377
[500] train's binary_logloss: 0.63434 valid's binary_logloss: 0.653884
Early stopping, best iteration is:
[488] train's binary_logloss: 0.63474 valid's binary_logloss: 0.653747
regularization_factors, val_score: 0.653747: 25%|##5 | 5/20 [00:05<00:18, 1.20s/it][I 2020-09-27 04:42:04,966] Trial 47 finished with value: 0.6537470788672778 and parameters: {'lambda_l1': 1.2446439098514144e-08, 'lambda_l2': 4.099446197874459e-06}. Best is trial 43 with value: 0.6537470788623818.
regularization_factors, val_score: 0.653747: 25%|##5 | 5/20 [00:05<00:18, 1.20s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004710 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657307 valid's binary_logloss: 0.665041
[200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725
[300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506
[400] train's binary_logloss: 0.637619 valid's binary_logloss: 0.654377
[500] train's binary_logloss: 0.63434 valid's binary_logloss: 0.653884
Early stopping, best iteration is:
[488] train's binary_logloss: 0.63474 valid's binary_logloss: 0.653747
regularization_factors, val_score: 0.653747: 30%|### | 6/20 [00:06<00:15, 1.11s/it][I 2020-09-27 04:42:05,863] Trial 48 finished with value: 0.6537470788771779 and parameters: {'lambda_l1': 1.2455751430953037e-08, 'lambda_l2': 2.1195700656837565e-06}. Best is trial 43 with value: 0.6537470788623818.
regularization_factors, val_score: 0.653747: 30%|### | 6/20 [00:06<00:15, 1.11s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000751 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657307 valid's binary_logloss: 0.665041
[200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725
[300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506
[400] train's binary_logloss: 0.637619 valid's binary_logloss: 0.654377
[500] train's binary_logloss: 0.63434 valid's binary_logloss: 0.653884
Early stopping, best iteration is:
[488] train's binary_logloss: 0.63474 valid's binary_logloss: 0.653747
regularization_factors, val_score: 0.653747: 35%|###5 | 7/20 [00:08<00:15, 1.18s/it][I 2020-09-27 04:42:07,198] Trial 49 finished with value: 0.6537470788676445 and parameters: {'lambda_l1': 1.3371676440125825e-08, 'lambda_l2': 4.133315482813227e-06}. Best is trial 43 with value: 0.6537470788623818.
regularization_factors, val_score: 0.653747: 35%|###5 | 7/20 [00:08<00:15, 1.18s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000868 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657307 valid's binary_logloss: 0.665041
[200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725
[300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506
[400] train's binary_logloss: 0.637619 valid's binary_logloss: 0.654377
[500] train's binary_logloss: 0.63434 valid's binary_logloss: 0.653884
Early stopping, best iteration is:
[488] train's binary_logloss: 0.63474 valid's binary_logloss: 0.653747
regularization_factors, val_score: 0.653747: 40%|#### | 8/20 [00:09<00:13, 1.14s/it][I 2020-09-27 04:42:08,255] Trial 50 finished with value: 0.653747078859269 and parameters: {'lambda_l1': 1.1210362012517112e-08, 'lambda_l2': 6.439200774621681e-06}. Best is trial 50 with value: 0.653747078859269.
regularization_factors, val_score: 0.653747: 40%|#### | 8/20 [00:09<00:13, 1.14s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000976 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657307 valid's binary_logloss: 0.665041
[200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725
[300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506
[400] train's binary_logloss: 0.637619 valid's binary_logloss: 0.654377
[500] train's binary_logloss: 0.63434 valid's binary_logloss: 0.653884
Early stopping, best iteration is:
[488] train's binary_logloss: 0.63474 valid's binary_logloss: 0.653747
regularization_factors, val_score: 0.653747: 45%|####5 | 9/20 [00:10<00:11, 1.07s/it][I 2020-09-27 04:42:09,153] Trial 51 finished with value: 0.6537470788593895 and parameters: {'lambda_l1': 1.3454643930345611e-08, 'lambda_l2': 6.467158854009067e-06}. Best is trial 50 with value: 0.653747078859269.
regularization_factors, val_score: 0.653747: 45%|####5 | 9/20 [00:10<00:11, 1.07s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.006696 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657307 valid's binary_logloss: 0.665041
[200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725
[300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506
[400] train's binary_logloss: 0.637619 valid's binary_logloss: 0.654377
[500] train's binary_logloss: 0.63434 valid's binary_logloss: 0.653884
Early stopping, best iteration is:
[488] train's binary_logloss: 0.63474 valid's binary_logloss: 0.653747
regularization_factors, val_score: 0.653747: 50%|##### | 10/20 [00:11<00:10, 1.02s/it][I 2020-09-27 04:42:10,075] Trial 52 finished with value: 0.6537470788591189 and parameters: {'lambda_l1': 1.1646094395679663e-08, 'lambda_l2': 6.358551933697898e-06}. Best is trial 52 with value: 0.6537470788591189.
regularization_factors, val_score: 0.653747: 50%|##### | 10/20 [00:11<00:10, 1.02s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011178 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657307 valid's binary_logloss: 0.665041
[200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725
[300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506
[400] train's binary_logloss: 0.637619 valid's binary_logloss: 0.654377
[500] train's binary_logloss: 0.63434 valid's binary_logloss: 0.653884
Early stopping, best iteration is:
[488] train's binary_logloss: 0.63474 valid's binary_logloss: 0.653747
regularization_factors, val_score: 0.653747: 55%|#####5 | 11/20 [00:12<00:09, 1.11s/it][I 2020-09-27 04:42:11,387] Trial 53 finished with value: 0.6537470788474831 and parameters: {'lambda_l1': 1.0924146750147604e-08, 'lambda_l2': 1.2976298424529484e-05}. Best is trial 53 with value: 0.6537470788474831.
regularization_factors, val_score: 0.653747: 55%|#####5 | 11/20 [00:12<00:09, 1.11s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005401 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657307 valid's binary_logloss: 0.665041
[200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725
[300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506
[400] train's binary_logloss: 0.637619 valid's binary_logloss: 0.654377
[500] train's binary_logloss: 0.63434 valid's binary_logloss: 0.653884
Early stopping, best iteration is:
[488] train's binary_logloss: 0.63474 valid's binary_logloss: 0.653747
regularization_factors, val_score: 0.653747: 60%|###### | 12/20 [00:13<00:08, 1.05s/it][I 2020-09-27 04:42:12,300] Trial 54 finished with value: 0.65374707865709 and parameters: {'lambda_l1': 1.1771590372611887e-08, 'lambda_l2': 6.880400470984478e-05}. Best is trial 54 with value: 0.65374707865709.
regularization_factors, val_score: 0.653747: 60%|###### | 12/20 [00:13<00:08, 1.05s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000845 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657307 valid's binary_logloss: 0.665041
[200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725
[300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506
[400] train's binary_logloss: 0.63762 valid's binary_logloss: 0.654377
[500] train's binary_logloss: 0.63434 valid's binary_logloss: 0.653884
Early stopping, best iteration is:
[488] train's binary_logloss: 0.634741 valid's binary_logloss: 0.653747
regularization_factors, val_score: 0.653747: 65%|######5 | 13/20 [00:14<00:07, 1.02s/it][I 2020-09-27 04:42:13,235] Trial 55 finished with value: 0.6537470737830029 and parameters: {'lambda_l1': 1.2904541521956938e-08, 'lambda_l2': 0.0015325466166573717}. Best is trial 55 with value: 0.6537470737830029.
regularization_factors, val_score: 0.653747: 65%|######5 | 13/20 [00:14<00:07, 1.02s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004533 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657308 valid's binary_logloss: 0.665041
[200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725
[300] train's binary_logloss: 0.64145 valid's binary_logloss: 0.65506
[400] train's binary_logloss: 0.637678 valid's binary_logloss: 0.654439
[500] train's binary_logloss: 0.634307 valid's binary_logloss: 0.653862
[600] train's binary_logloss: 0.631527 valid's binary_logloss: 0.653578
[700] train's binary_logloss: 0.628591 valid's binary_logloss: 0.653558
[800] train's binary_logloss: 0.625872 valid's binary_logloss: 0.653736
Early stopping, best iteration is:
[750] train's binary_logloss: 0.627225 valid's binary_logloss: 0.653307
regularization_factors, val_score: 0.653307: 70%|####### | 14/20 [00:15<00:07, 1.22s/it][I 2020-09-27 04:42:14,917] Trial 56 finished with value: 0.653306938112237 and parameters: {'lambda_l1': 1.1130628894602006e-05, 'lambda_l2': 0.007158975373652711}. Best is trial 56 with value: 0.653306938112237.
regularization_factors, val_score: 0.653307: 70%|####### | 14/20 [00:15<00:07, 1.22s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.019645 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657308 valid's binary_logloss: 0.665041
[200] train's binary_logloss: 0.646792 valid's binary_logloss: 0.657796
[300] train's binary_logloss: 0.641385 valid's binary_logloss: 0.655452
[400] train's binary_logloss: 0.637507 valid's binary_logloss: 0.654514
[500] train's binary_logloss: 0.634295 valid's binary_logloss: 0.653752
Early stopping, best iteration is:
[490] train's binary_logloss: 0.63465 valid's binary_logloss: 0.653564
regularization_factors, val_score: 0.653307: 75%|#######5 | 15/20 [00:16<00:05, 1.15s/it][I 2020-09-27 04:42:15,922] Trial 57 finished with value: 0.6535639999340161 and parameters: {'lambda_l1': 0.00026873318136680526, 'lambda_l2': 0.011949459080105706}. Best is trial 56 with value: 0.653306938112237.
regularization_factors, val_score: 0.653307: 75%|#######5 | 15/20 [00:16<00:05, 1.15s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000821 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657308 valid's binary_logloss: 0.665041
[200] train's binary_logloss: 0.646792 valid's binary_logloss: 0.657796
[300] train's binary_logloss: 0.641358 valid's binary_logloss: 0.655395
[400] train's binary_logloss: 0.637525 valid's binary_logloss: 0.654849
[500] train's binary_logloss: 0.63422 valid's binary_logloss: 0.654174
[600] train's binary_logloss: 0.631488 valid's binary_logloss: 0.654245
Early stopping, best iteration is:
[588] train's binary_logloss: 0.631831 valid's binary_logloss: 0.654014
regularization_factors, val_score: 0.653307: 80%|######## | 16/20 [00:17<00:04, 1.12s/it][I 2020-09-27 04:42:16,965] Trial 58 finished with value: 0.6540142566432265 and parameters: {'lambda_l1': 0.0002391919617921031, 'lambda_l2': 0.020184808500303845}. Best is trial 56 with value: 0.653306938112237.
regularization_factors, val_score: 0.653307: 80%|######## | 16/20 [00:17<00:04, 1.12s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004897 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657308 valid's binary_logloss: 0.665041
[200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725
[300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506
[400] train's binary_logloss: 0.63762 valid's binary_logloss: 0.654377
[500] train's binary_logloss: 0.634341 valid's binary_logloss: 0.653884
Early stopping, best iteration is:
[488] train's binary_logloss: 0.634741 valid's binary_logloss: 0.653747
regularization_factors, val_score: 0.653307: 85%|########5 | 17/20 [00:18<00:03, 1.06s/it][I 2020-09-27 04:42:17,888] Trial 59 finished with value: 0.6537470700910835 and parameters: {'lambda_l1': 0.00026680911329737025, 'lambda_l2': 0.0033110276633739268}. Best is trial 56 with value: 0.653306938112237.
regularization_factors, val_score: 0.653307: 85%|########5 | 17/20 [00:18<00:03, 1.06s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004895 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657308 valid's binary_logloss: 0.665041
[200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725
[300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506
[400] train's binary_logloss: 0.637621 valid's binary_logloss: 0.654377
[500] train's binary_logloss: 0.634341 valid's binary_logloss: 0.653884
Early stopping, best iteration is:
[488] train's binary_logloss: 0.634742 valid's binary_logloss: 0.653747
regularization_factors, val_score: 0.653307: 90%|######### | 18/20 [00:20<00:02, 1.17s/it][I 2020-09-27 04:42:19,323] Trial 60 finished with value: 0.6537470677716328 and parameters: {'lambda_l1': 0.0003014398725480866, 'lambda_l2': 0.004090143660649428}. Best is trial 56 with value: 0.653306938112237.
regularization_factors, val_score: 0.653307: 90%|######### | 18/20 [00:20<00:02, 1.17s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000797 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657308 valid's binary_logloss: 0.665041
[200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725
[300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506
[400] train's binary_logloss: 0.637621 valid's binary_logloss: 0.654377
[500] train's binary_logloss: 0.634342 valid's binary_logloss: 0.653884
Early stopping, best iteration is:
[488] train's binary_logloss: 0.634742 valid's binary_logloss: 0.653747
regularization_factors, val_score: 0.653307: 95%|#########5| 19/20 [00:21<00:01, 1.10s/it][I 2020-09-27 04:42:20,258] Trial 61 finished with value: 0.6537470687771331 and parameters: {'lambda_l1': 0.0005632008256227812, 'lambda_l2': 0.004446124070486416}. Best is trial 56 with value: 0.653306938112237.
regularization_factors, val_score: 0.653307: 95%|#########5| 19/20 [00:21<00:01, 1.10s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000827 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657308 valid's binary_logloss: 0.665041
[200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725
[300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506
[400] train's binary_logloss: 0.637621 valid's binary_logloss: 0.654377
[500] train's binary_logloss: 0.634341 valid's binary_logloss: 0.653884
Early stopping, best iteration is:
[488] train's binary_logloss: 0.634742 valid's binary_logloss: 0.653747
regularization_factors, val_score: 0.653307: 100%|##########| 20/20 [00:22<00:00, 1.04s/it][I 2020-09-27 04:42:21,158] Trial 62 finished with value: 0.6537470665554331 and parameters: {'lambda_l1': 0.0003132915633232416, 'lambda_l2': 0.004489563552243147}. Best is trial 56 with value: 0.653306938112237.
regularization_factors, val_score: 0.653307: 100%|##########| 20/20 [00:22<00:00, 1.11s/it]
min_data_in_leaf, val_score: 0.653307: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004952 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657308 valid's binary_logloss: 0.665041
[200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725
[300] train's binary_logloss: 0.641484 valid's binary_logloss: 0.655571
[400] train's binary_logloss: 0.63758 valid's binary_logloss: 0.654925
Early stopping, best iteration is:
[368] train's binary_logloss: 0.638713 valid's binary_logloss: 0.654841
min_data_in_leaf, val_score: 0.653307: 20%|## | 1/5 [00:00<00:03, 1.21it/s][I 2020-09-27 04:42:21,996] Trial 63 finished with value: 0.6548408998104529 and parameters: {'min_child_samples': 5}. Best is trial 63 with value: 0.6548408998104529.
min_data_in_leaf, val_score: 0.653307: 20%|## | 1/5 [00:00<00:03, 1.21it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005026 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657308 valid's binary_logloss: 0.665041
[200] train's binary_logloss: 0.646871 valid's binary_logloss: 0.657816
[300] train's binary_logloss: 0.641553 valid's binary_logloss: 0.655747
[400] train's binary_logloss: 0.637618 valid's binary_logloss: 0.655166
[500] train's binary_logloss: 0.63433 valid's binary_logloss: 0.654821
Early stopping, best iteration is:
[483] train's binary_logloss: 0.634862 valid's binary_logloss: 0.654649
min_data_in_leaf, val_score: 0.653307: 40%|#### | 2/5 [00:02<00:02, 1.04it/s][I 2020-09-27 04:42:23,270] Trial 64 finished with value: 0.6546487352338517 and parameters: {'min_child_samples': 25}. Best is trial 64 with value: 0.6546487352338517.
min_data_in_leaf, val_score: 0.653307: 40%|#### | 2/5 [00:02<00:02, 1.04it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004979 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657308 valid's binary_logloss: 0.665041
[200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725
[300] train's binary_logloss: 0.641452 valid's binary_logloss: 0.655218
[400] train's binary_logloss: 0.637601 valid's binary_logloss: 0.654481
[500] train's binary_logloss: 0.634142 valid's binary_logloss: 0.653967
[600] train's binary_logloss: 0.631213 valid's binary_logloss: 0.654248
Early stopping, best iteration is:
[539] train's binary_logloss: 0.632957 valid's binary_logloss: 0.653768
min_data_in_leaf, val_score: 0.653307: 60%|###### | 3/5 [00:03<00:01, 1.02it/s][I 2020-09-27 04:42:24,298] Trial 65 finished with value: 0.653767937805501 and parameters: {'min_child_samples': 10}. Best is trial 65 with value: 0.653767937805501.
min_data_in_leaf, val_score: 0.653307: 60%|###### | 3/5 [00:03<00:01, 1.02it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000881 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657308 valid's binary_logloss: 0.665041
[200] train's binary_logloss: 0.646871 valid's binary_logloss: 0.657816
[300] train's binary_logloss: 0.641452 valid's binary_logloss: 0.655693
[400] train's binary_logloss: 0.637718 valid's binary_logloss: 0.654576
[500] train's binary_logloss: 0.634473 valid's binary_logloss: 0.654212
[600] train's binary_logloss: 0.631693 valid's binary_logloss: 0.654007
Early stopping, best iteration is:
[580] train's binary_logloss: 0.63221 valid's binary_logloss: 0.653668
min_data_in_leaf, val_score: 0.653307: 80%|######## | 4/5 [00:04<00:01, 1.01s/it][I 2020-09-27 04:42:25,378] Trial 66 finished with value: 0.6536677236980378 and parameters: {'min_child_samples': 50}. Best is trial 66 with value: 0.6536677236980378.
min_data_in_leaf, val_score: 0.653307: 80%|######## | 4/5 [00:04<00:01, 1.01s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004998 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386
[LightGBM] [Info] Start training from score 0.022386
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657308 valid's binary_logloss: 0.665041
[200] train's binary_logloss: 0.646872 valid's binary_logloss: 0.657799
[300] train's binary_logloss: 0.641589 valid's binary_logloss: 0.655365
[400] train's binary_logloss: 0.638214 valid's binary_logloss: 0.65449
[500] train's binary_logloss: 0.635191 valid's binary_logloss: 0.653777
[600] train's binary_logloss: 0.632659 valid's binary_logloss: 0.653578
Early stopping, best iteration is:
[576] train's binary_logloss: 0.633154 valid's binary_logloss: 0.653401
min_data_in_leaf, val_score: 0.653307: 100%|##########| 5/5 [00:05<00:00, 1.17s/it][I 2020-09-27 04:42:26,912] Trial 67 finished with value: 0.653400524735626 and parameters: {'min_child_samples': 100}. Best is trial 67 with value: 0.653400524735626.
min_data_in_leaf, val_score: 0.653307: 100%|##########| 5/5 [00:05<00:00, 1.15s/it]
Fold : 4
[I 2020-09-27 04:42:26,996] A new study created in memory with name: no-name-95b73563-c9f0-498a-9add-fe0ff51242f2
feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000967 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.569629 valid's binary_logloss: 0.666525
Early stopping, best iteration is:
[63] train's binary_logloss: 0.595461 valid's binary_logloss: 0.665516
feature_fraction, val_score: 0.665516: 14%|#4 | 1/7 [00:00<00:03, 1.75it/s][I 2020-09-27 04:42:27,577] Trial 0 finished with value: 0.6655156235172205 and parameters: {'feature_fraction': 1.0}. Best is trial 0 with value: 0.6655156235172205.
feature_fraction, val_score: 0.665516: 14%|#4 | 1/7 [00:00<00:03, 1.75it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000965 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.572137 valid's binary_logloss: 0.663019
Early stopping, best iteration is:
[48] train's binary_logloss: 0.610212 valid's binary_logloss: 0.661166
feature_fraction, val_score: 0.661166: 29%|##8 | 2/7 [00:01<00:02, 1.85it/s][I 2020-09-27 04:42:28,047] Trial 1 finished with value: 0.6611663859231157 and parameters: {'feature_fraction': 0.8}. Best is trial 1 with value: 0.6611663859231157.
feature_fraction, val_score: 0.661166: 29%|##8 | 2/7 [00:01<00:02, 1.85it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004690 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.57535 valid's binary_logloss: 0.662198
Early stopping, best iteration is:
[58] train's binary_logloss: 0.604866 valid's binary_logloss: 0.660561
feature_fraction, val_score: 0.660561: 43%|####2 | 3/7 [00:01<00:02, 1.92it/s][I 2020-09-27 04:42:28,523] Trial 2 finished with value: 0.660560693860469 and parameters: {'feature_fraction': 0.6}. Best is trial 2 with value: 0.660560693860469.
feature_fraction, val_score: 0.660561: 43%|####2 | 3/7 [00:01<00:02, 1.92it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000938 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.571052 valid's binary_logloss: 0.6616
Early stopping, best iteration is:
[50] train's binary_logloss: 0.607427 valid's binary_logloss: 0.660867
feature_fraction, val_score: 0.660561: 57%|#####7 | 4/7 [00:02<00:01, 1.94it/s][I 2020-09-27 04:42:29,026] Trial 3 finished with value: 0.6608672257275293 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 2 with value: 0.660560693860469.
feature_fraction, val_score: 0.660561: 57%|#####7 | 4/7 [00:02<00:01, 1.94it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000868 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.574313 valid's binary_logloss: 0.662136
Early stopping, best iteration is:
[97] train's binary_logloss: 0.576201 valid's binary_logloss: 0.661927
feature_fraction, val_score: 0.660561: 71%|#######1 | 5/7 [00:02<00:01, 1.79it/s][I 2020-09-27 04:42:29,690] Trial 4 finished with value: 0.661927284753647 and parameters: {'feature_fraction': 0.7}. Best is trial 2 with value: 0.660560693860469.
feature_fraction, val_score: 0.660561: 71%|#######1 | 5/7 [00:02<00:01, 1.79it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012827 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.578468 valid's binary_logloss: 0.661585
Early stopping, best iteration is:
[54] train's binary_logloss: 0.610009 valid's binary_logloss: 0.661106
feature_fraction, val_score: 0.660561: 86%|########5 | 6/7 [00:03<00:00, 1.58it/s][I 2020-09-27 04:42:30,492] Trial 5 finished with value: 0.6611057071152201 and parameters: {'feature_fraction': 0.5}. Best is trial 2 with value: 0.660560693860469.
feature_fraction, val_score: 0.660561: 86%|########5 | 6/7 [00:03<00:00, 1.58it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000508 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.581603 valid's binary_logloss: 0.660601
Early stopping, best iteration is:
[81] train's binary_logloss: 0.593654 valid's binary_logloss: 0.660443
feature_fraction, val_score: 0.660443: 100%|##########| 7/7 [00:03<00:00, 1.68it/s][I 2020-09-27 04:42:30,996] Trial 6 finished with value: 0.6604425614247917 and parameters: {'feature_fraction': 0.4}. Best is trial 6 with value: 0.6604425614247917.
feature_fraction, val_score: 0.660443: 100%|##########| 7/7 [00:04<00:00, 1.75it/s]
num_leaves, val_score: 0.660443: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004658 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.356495 valid's binary_logloss: 0.671539
Early stopping, best iteration is:
[41] train's binary_logloss: 0.499664 valid's binary_logloss: 0.664498
num_leaves, val_score: 0.660443: 5%|5 | 1/20 [00:00<00:16, 1.14it/s][I 2020-09-27 04:42:31,893] Trial 7 finished with value: 0.6644984899613434 and parameters: {'num_leaves': 190}. Best is trial 7 with value: 0.6644984899613434.
num_leaves, val_score: 0.660443: 5%|5 | 1/20 [00:00<00:16, 1.14it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000609 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.526375 valid's binary_logloss: 0.665053
Early stopping, best iteration is:
[67] train's binary_logloss: 0.561219 valid's binary_logloss: 0.662832
num_leaves, val_score: 0.660443: 10%|# | 2/20 [00:01<00:14, 1.28it/s][I 2020-09-27 04:42:32,449] Trial 8 finished with value: 0.6628317798276503 and parameters: {'num_leaves': 60}. Best is trial 8 with value: 0.6628317798276503.
num_leaves, val_score: 0.660443: 10%|# | 2/20 [00:01<00:14, 1.28it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000379 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.608789 valid's binary_logloss: 0.659931
[200] train's binary_logloss: 0.573948 valid's binary_logloss: 0.661564
Early stopping, best iteration is:
[106] train's binary_logloss: 0.606284 valid's binary_logloss: 0.659493
num_leaves, val_score: 0.659493: 15%|#5 | 3/20 [00:01<00:11, 1.48it/s][I 2020-09-27 04:42:32,881] Trial 9 finished with value: 0.6594930122196023 and parameters: {'num_leaves': 19}. Best is trial 9 with value: 0.6594930122196023.
num_leaves, val_score: 0.659493: 15%|#5 | 3/20 [00:01<00:11, 1.48it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000374 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.62851 valid's binary_logloss: 0.661694
[200] train's binary_logloss: 0.606786 valid's binary_logloss: 0.660582
Early stopping, best iteration is:
[177] train's binary_logloss: 0.611571 valid's binary_logloss: 0.660135
num_leaves, val_score: 0.659493: 20%|## | 4/20 [00:02<00:10, 1.56it/s][I 2020-09-27 04:42:33,444] Trial 10 finished with value: 0.6601347890202464 and parameters: {'num_leaves': 11}. Best is trial 9 with value: 0.6594930122196023.
num_leaves, val_score: 0.659493: 20%|## | 4/20 [00:02<00:10, 1.56it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002220 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.623153 valid's binary_logloss: 0.659612
[200] train's binary_logloss: 0.598247 valid's binary_logloss: 0.660773
Early stopping, best iteration is:
[106] train's binary_logloss: 0.621362 valid's binary_logloss: 0.659491
num_leaves, val_score: 0.659491: 25%|##5 | 5/20 [00:03<00:10, 1.41it/s][I 2020-09-27 04:42:34,315] Trial 11 finished with value: 0.6594905167648981 and parameters: {'num_leaves': 13}. Best is trial 11 with value: 0.6594905167648981.
num_leaves, val_score: 0.659491: 25%|##5 | 5/20 [00:03<00:10, 1.41it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000240 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.633624 valid's binary_logloss: 0.661805
[200] train's binary_logloss: 0.6154 valid's binary_logloss: 0.661626
Early stopping, best iteration is:
[118] train's binary_logloss: 0.629667 valid's binary_logloss: 0.661053
num_leaves, val_score: 0.659491: 30%|### | 6/20 [00:03<00:08, 1.61it/s][I 2020-09-27 04:42:34,733] Trial 12 finished with value: 0.6610531437027433 and parameters: {'num_leaves': 9}. Best is trial 11 with value: 0.6594905167648981.
num_leaves, val_score: 0.659491: 30%|### | 6/20 [00:03<00:08, 1.61it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000443 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.489932 valid's binary_logloss: 0.664356
Early stopping, best iteration is:
[42] train's binary_logloss: 0.574496 valid's binary_logloss: 0.661322
num_leaves, val_score: 0.659491: 35%|###5 | 7/20 [00:04<00:07, 1.63it/s][I 2020-09-27 04:42:35,325] Trial 13 finished with value: 0.6613221980695407 and parameters: {'num_leaves': 83}. Best is trial 11 with value: 0.6594905167648981.
num_leaves, val_score: 0.659491: 35%|###5 | 7/20 [00:04<00:07, 1.63it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000456 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.668466 valid's binary_logloss: 0.674773
[200] train's binary_logloss: 0.658603 valid's binary_logloss: 0.667149
[300] train's binary_logloss: 0.653224 valid's binary_logloss: 0.663743
[400] train's binary_logloss: 0.649988 valid's binary_logloss: 0.661832
[500] train's binary_logloss: 0.647929 valid's binary_logloss: 0.661043
[600] train's binary_logloss: 0.646563 valid's binary_logloss: 0.660633
[700] train's binary_logloss: 0.645608 valid's binary_logloss: 0.660633
[800] train's binary_logloss: 0.644889 valid's binary_logloss: 0.660438
[900] train's binary_logloss: 0.644316 valid's binary_logloss: 0.66034
[1000] train's binary_logloss: 0.64384 valid's binary_logloss: 0.660316
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.64384 valid's binary_logloss: 0.660316
num_leaves, val_score: 0.659491: 40%|#### | 8/20 [00:05<00:09, 1.28it/s][I 2020-09-27 04:42:36,492] Trial 14 finished with value: 0.6603164467173235 and parameters: {'num_leaves': 2}. Best is trial 11 with value: 0.6594905167648981.
num_leaves, val_score: 0.659491: 40%|#### | 8/20 [00:05<00:09, 1.28it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000395 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.481541 valid's binary_logloss: 0.666981
Early stopping, best iteration is:
[44] train's binary_logloss: 0.566666 valid's binary_logloss: 0.662735
num_leaves, val_score: 0.659491: 45%|####5 | 9/20 [00:06<00:08, 1.35it/s][I 2020-09-27 04:42:37,143] Trial 15 finished with value: 0.6627351233501201 and parameters: {'num_leaves': 88}. Best is trial 11 with value: 0.6594905167648981.
num_leaves, val_score: 0.659491: 45%|####5 | 9/20 [00:06<00:08, 1.35it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000349 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.3973 valid's binary_logloss: 0.672802
Early stopping, best iteration is:
[29] train's binary_logloss: 0.560753 valid's binary_logloss: 0.66553
num_leaves, val_score: 0.659491: 50%|##### | 10/20 [00:07<00:09, 1.08it/s][I 2020-09-27 04:42:38,498] Trial 16 finished with value: 0.6655299653769075 and parameters: {'num_leaves': 152}. Best is trial 11 with value: 0.6594905167648981.
num_leaves, val_score: 0.659491: 50%|##### | 10/20 [00:07<00:09, 1.08it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000373 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.299677 valid's binary_logloss: 0.679053
Early stopping, best iteration is:
[28] train's binary_logloss: 0.514613 valid's binary_logloss: 0.667796
num_leaves, val_score: 0.659491: 55%|#####5 | 11/20 [00:08<00:08, 1.03it/s][I 2020-09-27 04:42:39,559] Trial 17 finished with value: 0.6677957520143049 and parameters: {'num_leaves': 256}. Best is trial 11 with value: 0.6594905167648981.
num_leaves, val_score: 0.659491: 55%|#####5 | 11/20 [00:08<00:08, 1.03it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000315 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.5506 valid's binary_logloss: 0.661771
Early stopping, best iteration is:
[46] train's binary_logloss: 0.602349 valid's binary_logloss: 0.660582
num_leaves, val_score: 0.659491: 60%|###### | 12/20 [00:08<00:06, 1.23it/s][I 2020-09-27 04:42:40,010] Trial 18 finished with value: 0.6605823886498043 and parameters: {'num_leaves': 47}. Best is trial 11 with value: 0.6594905167648981.
num_leaves, val_score: 0.659491: 60%|###### | 12/20 [00:09<00:06, 1.23it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000360 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.573596 valid's binary_logloss: 0.659418
Early stopping, best iteration is:
[95] train's binary_logloss: 0.576854 valid's binary_logloss: 0.658933
num_leaves, val_score: 0.658933: 65%|######5 | 13/20 [00:09<00:04, 1.40it/s][I 2020-09-27 04:42:40,490] Trial 19 finished with value: 0.6589331658992789 and parameters: {'num_leaves': 35}. Best is trial 19 with value: 0.6589331658992789.
num_leaves, val_score: 0.658933: 65%|######5 | 13/20 [00:09<00:04, 1.40it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000370 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.433641 valid's binary_logloss: 0.669498
Early stopping, best iteration is:
[30] train's binary_logloss: 0.5737 valid's binary_logloss: 0.666057
num_leaves, val_score: 0.658933: 70%|####### | 14/20 [00:10<00:04, 1.31it/s][I 2020-09-27 04:42:41,370] Trial 20 finished with value: 0.666056567776029 and parameters: {'num_leaves': 123}. Best is trial 19 with value: 0.6589331658992789.
num_leaves, val_score: 0.658933: 70%|####### | 14/20 [00:10<00:04, 1.31it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011203 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.578104 valid's binary_logloss: 0.662017
Early stopping, best iteration is:
[72] train's binary_logloss: 0.596911 valid's binary_logloss: 0.661574
num_leaves, val_score: 0.658933: 75%|#######5 | 15/20 [00:10<00:03, 1.41it/s][I 2020-09-27 04:42:41,958] Trial 21 finished with value: 0.6615737018014829 and parameters: {'num_leaves': 33}. Best is trial 19 with value: 0.6589331658992789.
num_leaves, val_score: 0.658933: 75%|#######5 | 15/20 [00:10<00:03, 1.41it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000381 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.668466 valid's binary_logloss: 0.674773
[200] train's binary_logloss: 0.658603 valid's binary_logloss: 0.667149
[300] train's binary_logloss: 0.653224 valid's binary_logloss: 0.663743
[400] train's binary_logloss: 0.649988 valid's binary_logloss: 0.661832
[500] train's binary_logloss: 0.647929 valid's binary_logloss: 0.661043
[600] train's binary_logloss: 0.646563 valid's binary_logloss: 0.660633
[700] train's binary_logloss: 0.645608 valid's binary_logloss: 0.660633
[800] train's binary_logloss: 0.644889 valid's binary_logloss: 0.660438
[900] train's binary_logloss: 0.644316 valid's binary_logloss: 0.66034
[1000] train's binary_logloss: 0.64384 valid's binary_logloss: 0.660316
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.64384 valid's binary_logloss: 0.660316
num_leaves, val_score: 0.658933: 80%|######## | 16/20 [00:12<00:03, 1.18it/s][I 2020-09-27 04:42:43,132] Trial 22 finished with value: 0.6603164467173235 and parameters: {'num_leaves': 2}. Best is trial 19 with value: 0.6589331658992789.
num_leaves, val_score: 0.658933: 80%|######## | 16/20 [00:12<00:03, 1.18it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000388 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.581603 valid's binary_logloss: 0.660601
Early stopping, best iteration is:
[81] train's binary_logloss: 0.593654 valid's binary_logloss: 0.660443
num_leaves, val_score: 0.658933: 85%|########5 | 17/20 [00:12<00:02, 1.34it/s][I 2020-09-27 04:42:43,636] Trial 23 finished with value: 0.6604425614247916 and parameters: {'num_leaves': 31}. Best is trial 19 with value: 0.6589331658992789.
num_leaves, val_score: 0.658933: 85%|########5 | 17/20 [00:12<00:02, 1.34it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000430 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.517038 valid's binary_logloss: 0.664708
Early stopping, best iteration is:
[42] train's binary_logloss: 0.589858 valid's binary_logloss: 0.660965
num_leaves, val_score: 0.658933: 90%|######### | 18/20 [00:13<00:01, 1.47it/s][I 2020-09-27 04:42:44,171] Trial 24 finished with value: 0.660965322744683 and parameters: {'num_leaves': 66}. Best is trial 19 with value: 0.6589331658992789.
num_leaves, val_score: 0.658933: 90%|######### | 18/20 [00:13<00:01, 1.47it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000404 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.590111 valid's binary_logloss: 0.659776
Early stopping, best iteration is:
[96] train's binary_logloss: 0.592038 valid's binary_logloss: 0.65953
num_leaves, val_score: 0.658933: 95%|#########5| 19/20 [00:13<00:00, 1.61it/s][I 2020-09-27 04:42:44,655] Trial 25 finished with value: 0.6595301896319116 and parameters: {'num_leaves': 27}. Best is trial 19 with value: 0.6589331658992789.
num_leaves, val_score: 0.658933: 95%|#########5| 19/20 [00:13<00:00, 1.61it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000349 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.454148 valid's binary_logloss: 0.667807
Early stopping, best iteration is:
[45] train's binary_logloss: 0.54925 valid's binary_logloss: 0.664334
num_leaves, val_score: 0.658933: 100%|##########| 20/20 [00:14<00:00, 1.28it/s][I 2020-09-27 04:42:45,813] Trial 26 finished with value: 0.6643342584470427 and parameters: {'num_leaves': 106}. Best is trial 19 with value: 0.6589331658992789.
num_leaves, val_score: 0.658933: 100%|##########| 20/20 [00:14<00:00, 1.35it/s]
bagging, val_score: 0.658933: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000378 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.57621 valid's binary_logloss: 0.663844
Early stopping, best iteration is:
[82] train's binary_logloss: 0.587864 valid's binary_logloss: 0.662968
bagging, val_score: 0.658933: 10%|# | 1/10 [00:00<00:04, 1.88it/s][I 2020-09-27 04:42:46,361] Trial 27 finished with value: 0.6629683704520333 and parameters: {'bagging_fraction': 0.6524827037560431, 'bagging_freq': 6}. Best is trial 27 with value: 0.6629683704520333.
bagging, val_score: 0.658933: 10%|# | 1/10 [00:00<00:04, 1.88it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000416 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.573429 valid's binary_logloss: 0.66344
Early stopping, best iteration is:
[70] train's binary_logloss: 0.595186 valid's binary_logloss: 0.661013
bagging, val_score: 0.658933: 20%|## | 2/10 [00:00<00:04, 1.96it/s][I 2020-09-27 04:42:46,824] Trial 28 finished with value: 0.6610133814310396 and parameters: {'bagging_fraction': 0.9909619380758035, 'bagging_freq': 1}. Best is trial 28 with value: 0.6610133814310396.
bagging, val_score: 0.658933: 20%|## | 2/10 [00:01<00:04, 1.96it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000474 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.580089 valid's binary_logloss: 0.668617
Early stopping, best iteration is:
[39] train's binary_logloss: 0.625728 valid's binary_logloss: 0.665767
bagging, val_score: 0.658933: 30%|### | 3/10 [00:01<00:03, 2.09it/s][I 2020-09-27 04:42:47,229] Trial 29 finished with value: 0.6657670269268177 and parameters: {'bagging_fraction': 0.42763463150211456, 'bagging_freq': 1}. Best is trial 28 with value: 0.6610133814310396.
bagging, val_score: 0.658933: 30%|### | 3/10 [00:01<00:03, 2.09it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000410 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.573775 valid's binary_logloss: 0.662206
Early stopping, best iteration is:
[84] train's binary_logloss: 0.585244 valid's binary_logloss: 0.661816
bagging, val_score: 0.658933: 40%|#### | 4/10 [00:01<00:02, 2.03it/s][I 2020-09-27 04:42:47,755] Trial 30 finished with value: 0.6618156054088046 and parameters: {'bagging_fraction': 0.9647411027897097, 'bagging_freq': 7}. Best is trial 28 with value: 0.6610133814310396.
bagging, val_score: 0.658933: 40%|#### | 4/10 [00:01<00:02, 2.03it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004478 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.579254 valid's binary_logloss: 0.665041
Early stopping, best iteration is:
[60] train's binary_logloss: 0.607077 valid's binary_logloss: 0.658928
bagging, val_score: 0.658928: 50%|##### | 5/10 [00:02<00:02, 2.12it/s][I 2020-09-27 04:42:48,176] Trial 31 finished with value: 0.6589276849000186 and parameters: {'bagging_fraction': 0.4607699724160956, 'bagging_freq': 4}. Best is trial 31 with value: 0.6589276849000186.
bagging, val_score: 0.658928: 50%|##### | 5/10 [00:02<00:02, 2.12it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000381 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.580916 valid's binary_logloss: 0.671067
Early stopping, best iteration is:
[36] train's binary_logloss: 0.628471 valid's binary_logloss: 0.662306
bagging, val_score: 0.658928: 60%|###### | 6/10 [00:03<00:02, 1.82it/s][I 2020-09-27 04:42:48,914] Trial 32 finished with value: 0.6623061121182913 and parameters: {'bagging_fraction': 0.4089026397973168, 'bagging_freq': 4}. Best is trial 31 with value: 0.6589276849000186.
bagging, val_score: 0.658928: 60%|###### | 6/10 [00:03<00:02, 1.82it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001551 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.57696 valid's binary_logloss: 0.664792
Early stopping, best iteration is:
[44] train's binary_logloss: 0.620171 valid's binary_logloss: 0.661311
bagging, val_score: 0.658928: 70%|####### | 7/10 [00:03<00:01, 1.75it/s][I 2020-09-27 04:42:49,530] Trial 33 finished with value: 0.6613113079015099 and parameters: {'bagging_fraction': 0.5861812306541639, 'bagging_freq': 4}. Best is trial 31 with value: 0.6589276849000186.
bagging, val_score: 0.658928: 70%|####### | 7/10 [00:03<00:01, 1.75it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000437 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.575286 valid's binary_logloss: 0.663569
Early stopping, best iteration is:
[62] train's binary_logloss: 0.60231 valid's binary_logloss: 0.661403
bagging, val_score: 0.658928: 80%|######## | 8/10 [00:04<00:01, 1.80it/s][I 2020-09-27 04:42:50,048] Trial 34 finished with value: 0.6614028699158252 and parameters: {'bagging_fraction': 0.8070642452561286, 'bagging_freq': 3}. Best is trial 31 with value: 0.6589276849000186.
bagging, val_score: 0.658928: 80%|######## | 8/10 [00:04<00:01, 1.80it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000238 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.578059 valid's binary_logloss: 0.66561
Early stopping, best iteration is:
[57] train's binary_logloss: 0.60883 valid's binary_logloss: 0.660744
bagging, val_score: 0.658928: 90%|######### | 9/10 [00:04<00:00, 1.85it/s][I 2020-09-27 04:42:50,558] Trial 35 finished with value: 0.6607440151387087 and parameters: {'bagging_fraction': 0.5229694779889069, 'bagging_freq': 5}. Best is trial 31 with value: 0.6589276849000186.
bagging, val_score: 0.658928: 90%|######### | 9/10 [00:04<00:00, 1.85it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000982 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.574809 valid's binary_logloss: 0.665725
Early stopping, best iteration is:
[73] train's binary_logloss: 0.593617 valid's binary_logloss: 0.664242
bagging, val_score: 0.658928: 100%|##########| 10/10 [00:05<00:00, 1.76it/s][I 2020-09-27 04:42:51,184] Trial 36 finished with value: 0.6642418801252681 and parameters: {'bagging_fraction': 0.7857998781191295, 'bagging_freq': 2}. Best is trial 31 with value: 0.6589276849000186.
bagging, val_score: 0.658928: 100%|##########| 10/10 [00:05<00:00, 1.86it/s]
feature_fraction_stage2, val_score: 0.658928: 0%| | 0/3 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004751 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.576158 valid's binary_logloss: 0.667365
Early stopping, best iteration is:
[40] train's binary_logloss: 0.622262 valid's binary_logloss: 0.662845
feature_fraction_stage2, val_score: 0.658928: 33%|###3 | 1/3 [00:00<00:00, 2.32it/s][I 2020-09-27 04:42:51,633] Trial 37 finished with value: 0.6628451796786387 and parameters: {'feature_fraction': 0.44800000000000006}. Best is trial 37 with value: 0.6628451796786387.
feature_fraction_stage2, val_score: 0.658928: 33%|###3 | 1/3 [00:00<00:00, 2.32it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000343 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.577968 valid's binary_logloss: 0.668202
Early stopping, best iteration is:
[34] train's binary_logloss: 0.629452 valid's binary_logloss: 0.662397
feature_fraction_stage2, val_score: 0.658928: 67%|######6 | 2/3 [00:00<00:00, 2.34it/s][I 2020-09-27 04:42:52,051] Trial 38 finished with value: 0.6623967977826815 and parameters: {'feature_fraction': 0.41600000000000004}. Best is trial 38 with value: 0.6623967977826815.
feature_fraction_stage2, val_score: 0.658928: 67%|######6 | 2/3 [00:00<00:00, 2.34it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007814 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.576158 valid's binary_logloss: 0.667365
Early stopping, best iteration is:
[40] train's binary_logloss: 0.622262 valid's binary_logloss: 0.662845
feature_fraction_stage2, val_score: 0.658928: 100%|##########| 3/3 [00:01<00:00, 2.00it/s][I 2020-09-27 04:42:52,719] Trial 39 finished with value: 0.6628451796786387 and parameters: {'feature_fraction': 0.48000000000000004}. Best is trial 38 with value: 0.6623967977826815.
feature_fraction_stage2, val_score: 0.658928: 100%|##########| 3/3 [00:01<00:00, 1.97it/s]
regularization_factors, val_score: 0.658928: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016489 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.578756 valid's binary_logloss: 0.666904
Early stopping, best iteration is:
[60] train's binary_logloss: 0.606361 valid's binary_logloss: 0.662873
regularization_factors, val_score: 0.658928: 5%|5 | 1/20 [00:00<00:12, 1.52it/s][I 2020-09-27 04:42:53,400] Trial 40 finished with value: 0.6628730879769316 and parameters: {'lambda_l1': 0.00027860521966329625, 'lambda_l2': 0.23171527606664835}. Best is trial 40 with value: 0.6628730879769316.
regularization_factors, val_score: 0.658928: 5%|5 | 1/20 [00:00<00:12, 1.52it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000377 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.57899 valid's binary_logloss: 0.664607
Early stopping, best iteration is:
[60] train's binary_logloss: 0.607074 valid's binary_logloss: 0.658905
regularization_factors, val_score: 0.658905: 10%|# | 2/20 [00:01<00:11, 1.62it/s][I 2020-09-27 04:42:53,923] Trial 41 finished with value: 0.6589045464246605 and parameters: {'lambda_l1': 1.0562160571745345e-08, 'lambda_l2': 1.2335861617943539e-08}. Best is trial 41 with value: 0.6589045464246605.
regularization_factors, val_score: 0.658905: 10%|# | 2/20 [00:01<00:11, 1.62it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000451 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.579401 valid's binary_logloss: 0.664539
Early stopping, best iteration is:
[60] train's binary_logloss: 0.607081 valid's binary_logloss: 0.658927
regularization_factors, val_score: 0.658905: 15%|#5 | 3/20 [00:01<00:09, 1.70it/s][I 2020-09-27 04:42:54,440] Trial 42 finished with value: 0.6589271408352281 and parameters: {'lambda_l1': 1.4846795215967304e-08, 'lambda_l2': 2.1890701373620828e-08}. Best is trial 41 with value: 0.6589045464246605.
regularization_factors, val_score: 0.658905: 15%|#5 | 3/20 [00:01<00:09, 1.70it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000474 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.579252 valid's binary_logloss: 0.665028
Early stopping, best iteration is:
[60] train's binary_logloss: 0.607075 valid's binary_logloss: 0.658927
regularization_factors, val_score: 0.658905: 20%|## | 4/20 [00:02<00:09, 1.71it/s][I 2020-09-27 04:42:55,018] Trial 43 finished with value: 0.6589273467253626 and parameters: {'lambda_l1': 1.1433697445980986e-08, 'lambda_l2': 1.3023515499807511e-08}. Best is trial 41 with value: 0.6589045464246605.
regularization_factors, val_score: 0.658905: 20%|## | 4/20 [00:02<00:09, 1.71it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000451 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.579216 valid's binary_logloss: 0.664431
Early stopping, best iteration is:
[60] train's binary_logloss: 0.607079 valid's binary_logloss: 0.658949
regularization_factors, val_score: 0.658905: 25%|##5 | 5/20 [00:02<00:08, 1.75it/s][I 2020-09-27 04:42:55,554] Trial 44 finished with value: 0.6589487630970773 and parameters: {'lambda_l1': 1.0140056425681878e-08, 'lambda_l2': 1.2300058074804937e-08}. Best is trial 41 with value: 0.6589045464246605.
regularization_factors, val_score: 0.658905: 25%|##5 | 5/20 [00:02<00:08, 1.75it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000479 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.578759 valid's binary_logloss: 0.665688
Early stopping, best iteration is:
[60] train's binary_logloss: 0.607073 valid's binary_logloss: 0.658904
regularization_factors, val_score: 0.658904: 30%|### | 6/20 [00:03<00:07, 1.84it/s][I 2020-09-27 04:42:56,039] Trial 45 finished with value: 0.658904002539462 and parameters: {'lambda_l1': 1.0953495812808097e-08, 'lambda_l2': 1.828459432273254e-08}. Best is trial 45 with value: 0.658904002539462.
regularization_factors, val_score: 0.658904: 30%|### | 6/20 [00:03<00:07, 1.84it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000378 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.579237 valid's binary_logloss: 0.665024
Early stopping, best iteration is:
[60] train's binary_logloss: 0.60708 valid's binary_logloss: 0.658928
regularization_factors, val_score: 0.658904: 35%|###5 | 7/20 [00:04<00:08, 1.49it/s][I 2020-09-27 04:42:57,006] Trial 46 finished with value: 0.6589280313288517 and parameters: {'lambda_l1': 1.0619074780064361e-08, 'lambda_l2': 1.470079266120941e-08}. Best is trial 45 with value: 0.658904002539462.
regularization_factors, val_score: 0.658904: 35%|###5 | 7/20 [00:04<00:08, 1.49it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000339 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.5794 valid's binary_logloss: 0.66454
Early stopping, best iteration is:
[60] train's binary_logloss: 0.607078 valid's binary_logloss: 0.658904
regularization_factors, val_score: 0.658904: 40%|#### | 8/20 [00:04<00:07, 1.63it/s][I 2020-09-27 04:42:57,486] Trial 47 finished with value: 0.6589040025411772 and parameters: {'lambda_l1': 1.3254321926524086e-08, 'lambda_l2': 2.193885866761881e-08}. Best is trial 45 with value: 0.658904002539462.
regularization_factors, val_score: 0.658904: 40%|#### | 8/20 [00:04<00:07, 1.63it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000424 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.579223 valid's binary_logloss: 0.664396
Early stopping, best iteration is:
[60] train's binary_logloss: 0.607078 valid's binary_logloss: 0.658927
regularization_factors, val_score: 0.658904: 45%|####5 | 9/20 [00:05<00:06, 1.76it/s][I 2020-09-27 04:42:57,951] Trial 48 finished with value: 0.6589267944230356 and parameters: {'lambda_l1': 2.4619075323767072e-08, 'lambda_l2': 2.017164260358036e-08}. Best is trial 45 with value: 0.658904002539462.
regularization_factors, val_score: 0.658904: 45%|####5 | 9/20 [00:05<00:06, 1.76it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000536 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.57921 valid's binary_logloss: 0.664407
Early stopping, best iteration is:
[60] train's binary_logloss: 0.607077 valid's binary_logloss: 0.658925
regularization_factors, val_score: 0.658904: 50%|##### | 10/20 [00:05<00:05, 1.81it/s][I 2020-09-27 04:42:58,460] Trial 49 finished with value: 0.6589253929659314 and parameters: {'lambda_l1': 1.2653815920250245e-08, 'lambda_l2': 2.3139792938926494e-08}. Best is trial 45 with value: 0.658904002539462.
regularization_factors, val_score: 0.658904: 50%|##### | 10/20 [00:05<00:05, 1.81it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004484 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.579211 valid's binary_logloss: 0.664425
Early stopping, best iteration is:
[60] train's binary_logloss: 0.607081 valid's binary_logloss: 0.658927
regularization_factors, val_score: 0.658904: 55%|#####5 | 11/20 [00:06<00:04, 1.90it/s][I 2020-09-27 04:42:58,929] Trial 50 finished with value: 0.6589271408099914 and parameters: {'lambda_l1': 1.3776160562135245e-08, 'lambda_l2': 2.6228878428660015e-07}. Best is trial 45 with value: 0.658904002539462.
regularization_factors, val_score: 0.658904: 55%|#####5 | 11/20 [00:06<00:04, 1.90it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000373 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.57921 valid's binary_logloss: 0.664409
Early stopping, best iteration is:
[60] train's binary_logloss: 0.607081 valid's binary_logloss: 0.658927
regularization_factors, val_score: 0.658904: 60%|###### | 12/20 [00:06<00:04, 1.94it/s][I 2020-09-27 04:42:59,423] Trial 51 finished with value: 0.658927140833919 and parameters: {'lambda_l1': 1.664666646319637e-08, 'lambda_l2': 4.250351022086274e-08}. Best is trial 45 with value: 0.658904002539462.
regularization_factors, val_score: 0.658904: 60%|###### | 12/20 [00:06<00:04, 1.94it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000390 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.57924 valid's binary_logloss: 0.665044
Early stopping, best iteration is:
[60] train's binary_logloss: 0.607082 valid's binary_logloss: 0.65895
regularization_factors, val_score: 0.658904: 65%|######5 | 13/20 [00:07<00:03, 1.94it/s][I 2020-09-27 04:42:59,937] Trial 52 finished with value: 0.6589497680473425 and parameters: {'lambda_l1': 1.439041833170667e-08, 'lambda_l2': 3.7007408240165166e-07}. Best is trial 45 with value: 0.658904002539462.
regularization_factors, val_score: 0.658904: 65%|######5 | 13/20 [00:07<00:03, 1.94it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000634 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.579255 valid's binary_logloss: 0.665048
Early stopping, best iteration is:
[60] train's binary_logloss: 0.607079 valid's binary_logloss: 0.658949
regularization_factors, val_score: 0.658904: 70%|####### | 14/20 [00:08<00:04, 1.47it/s][I 2020-09-27 04:43:01,006] Trial 53 finished with value: 0.6589494209805504 and parameters: {'lambda_l1': 1.7572183872374385e-07, 'lambda_l2': 9.82651046880018e-07}. Best is trial 45 with value: 0.658904002539462.
regularization_factors, val_score: 0.658904: 70%|####### | 14/20 [00:08<00:04, 1.47it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000481 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.579214 valid's binary_logloss: 0.664386
Early stopping, best iteration is:
[60] train's binary_logloss: 0.607075 valid's binary_logloss: 0.658904
regularization_factors, val_score: 0.658904: 75%|#######5 | 15/20 [00:08<00:03, 1.60it/s][I 2020-09-27 04:43:01,494] Trial 54 finished with value: 0.6589036559607917 and parameters: {'lambda_l1': 8.538730576389241e-07, 'lambda_l2': 1.1721978821541268e-08}. Best is trial 54 with value: 0.6589036559607917.
regularization_factors, val_score: 0.658904: 75%|#######5 | 15/20 [00:08<00:03, 1.60it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000414 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.579216 valid's binary_logloss: 0.664476
Early stopping, best iteration is:
[60] train's binary_logloss: 0.607081 valid's binary_logloss: 0.658927
regularization_factors, val_score: 0.658904: 80%|######## | 16/20 [00:09<00:02, 1.72it/s][I 2020-09-27 04:43:01,976] Trial 55 finished with value: 0.6589271405882069 and parameters: {'lambda_l1': 2.1323929878136295e-06, 'lambda_l2': 1.2914444598331435e-08}. Best is trial 54 with value: 0.6589036559607917.
regularization_factors, val_score: 0.658904: 80%|######## | 16/20 [00:09<00:02, 1.72it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004379 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.579258 valid's binary_logloss: 0.665048
Early stopping, best iteration is:
[60] train's binary_logloss: 0.607079 valid's binary_logloss: 0.658949
regularization_factors, val_score: 0.658904: 85%|########5 | 17/20 [00:09<00:01, 1.82it/s][I 2020-09-27 04:43:02,455] Trial 56 finished with value: 0.6589494210443969 and parameters: {'lambda_l1': 7.412222130813083e-07, 'lambda_l2': 1.1296738167689051e-08}. Best is trial 54 with value: 0.6589036559607917.
regularization_factors, val_score: 0.658904: 85%|########5 | 17/20 [00:09<00:01, 1.82it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000368 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.579256 valid's binary_logloss: 0.665048
Early stopping, best iteration is:
[60] train's binary_logloss: 0.607077 valid's binary_logloss: 0.658949
regularization_factors, val_score: 0.658904: 90%|######### | 18/20 [00:10<00:01, 1.87it/s][I 2020-09-27 04:43:02,951] Trial 57 finished with value: 0.6589493135745349 and parameters: {'lambda_l1': 2.9906109268867005e-07, 'lambda_l2': 1.0924154779547353e-05}. Best is trial 54 with value: 0.6589036559607917.
regularization_factors, val_score: 0.658904: 90%|######### | 18/20 [00:10<00:01, 1.87it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000413 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.579254 valid's binary_logloss: 0.665027
Early stopping, best iteration is:
[60] train's binary_logloss: 0.607078 valid's binary_logloss: 0.658927
regularization_factors, val_score: 0.658904: 95%|#########5| 19/20 [00:10<00:00, 1.95it/s][I 2020-09-27 04:43:03,413] Trial 58 finished with value: 0.658926859627446 and parameters: {'lambda_l1': 1.208558071943979e-07, 'lambda_l2': 1.4153597871067393e-07}. Best is trial 54 with value: 0.6589036559607917.
regularization_factors, val_score: 0.658904: 95%|#########5| 19/20 [00:10<00:00, 1.95it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004727 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.579258 valid's binary_logloss: 0.665048
Early stopping, best iteration is:
[60] train's binary_logloss: 0.607079 valid's binary_logloss: 0.658949
regularization_factors, val_score: 0.658904: 100%|##########| 20/20 [00:11<00:00, 1.94it/s][I 2020-09-27 04:43:03,937] Trial 59 finished with value: 0.6589494193426407 and parameters: {'lambda_l1': 1.386319739479547e-05, 'lambda_l2': 1.070237025489994e-08}. Best is trial 54 with value: 0.6589036559607917.
regularization_factors, val_score: 0.658904: 100%|##########| 20/20 [00:11<00:00, 1.78it/s]
min_data_in_leaf, val_score: 0.658904: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000392 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.576809 valid's binary_logloss: 0.668517
Early stopping, best iteration is:
[60] train's binary_logloss: 0.605306 valid's binary_logloss: 0.663449
min_data_in_leaf, val_score: 0.658904: 20%|## | 1/5 [00:00<00:03, 1.01it/s][I 2020-09-27 04:43:04,955] Trial 60 finished with value: 0.6634487376945314 and parameters: {'min_child_samples': 5}. Best is trial 60 with value: 0.6634487376945314.
min_data_in_leaf, val_score: 0.658904: 20%|## | 1/5 [00:01<00:03, 1.01it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000379 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.579781 valid's binary_logloss: 0.664717
Early stopping, best iteration is:
[60] train's binary_logloss: 0.607283 valid's binary_logloss: 0.661178
min_data_in_leaf, val_score: 0.658904: 40%|#### | 2/5 [00:01<00:02, 1.17it/s][I 2020-09-27 04:43:05,479] Trial 61 finished with value: 0.6611778602130459 and parameters: {'min_child_samples': 25}. Best is trial 61 with value: 0.6611778602130459.
min_data_in_leaf, val_score: 0.658904: 40%|#### | 2/5 [00:01<00:02, 1.17it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004590 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.587755 valid's binary_logloss: 0.665012
Early stopping, best iteration is:
[66] train's binary_logloss: 0.607178 valid's binary_logloss: 0.662389
min_data_in_leaf, val_score: 0.658904: 60%|###### | 3/5 [00:02<00:01, 1.34it/s][I 2020-09-27 04:43:05,977] Trial 62 finished with value: 0.6623889382406033 and parameters: {'min_child_samples': 100}. Best is trial 61 with value: 0.6611778602130459.
min_data_in_leaf, val_score: 0.658904: 60%|###### | 3/5 [00:02<00:01, 1.34it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000482 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.577819 valid's binary_logloss: 0.667336
Early stopping, best iteration is:
[60] train's binary_logloss: 0.605667 valid's binary_logloss: 0.663345
min_data_in_leaf, val_score: 0.658904: 80%|######## | 4/5 [00:02<00:00, 1.50it/s][I 2020-09-27 04:43:06,461] Trial 63 finished with value: 0.6633454558808694 and parameters: {'min_child_samples': 10}. Best is trial 61 with value: 0.6611778602130459.
min_data_in_leaf, val_score: 0.658904: 80%|######## | 4/5 [00:02<00:00, 1.50it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000449 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925
[LightGBM] [Info] Start training from score 0.023925
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.583153 valid's binary_logloss: 0.66795
Early stopping, best iteration is:
[59] train's binary_logloss: 0.610418 valid's binary_logloss: 0.663074
min_data_in_leaf, val_score: 0.658904: 100%|##########| 5/5 [00:02<00:00, 1.63it/s][I 2020-09-27 04:43:06,953] Trial 64 finished with value: 0.6630735000484437 and parameters: {'min_child_samples': 50}. Best is trial 61 with value: 0.6611778602130459.
min_data_in_leaf, val_score: 0.658904: 100%|##########| 5/5 [00:02<00:00, 1.67it/s]
Fold : 5
[I 2020-09-27 04:43:06,989] A new study created in memory with name: no-name-b151d15a-5a47-42dc-ace0-4fe96bb3c92c
feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004614 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.577114 valid's binary_logloss: 0.654088
Early stopping, best iteration is:
[60] train's binary_logloss: 0.604064 valid's binary_logloss: 0.652612
feature_fraction, val_score: 0.652612: 14%|#4 | 1/7 [00:00<00:02, 2.09it/s][I 2020-09-27 04:43:07,480] Trial 0 finished with value: 0.6526118300374438 and parameters: {'feature_fraction': 0.6}. Best is trial 0 with value: 0.6526118300374438.
feature_fraction, val_score: 0.652612: 14%|#4 | 1/7 [00:00<00:02, 2.09it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001021 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.570555 valid's binary_logloss: 0.652172
[200] train's binary_logloss: 0.514477 valid's binary_logloss: 0.655545
Early stopping, best iteration is:
[114] train's binary_logloss: 0.561489 valid's binary_logloss: 0.651258
feature_fraction, val_score: 0.651258: 29%|##8 | 2/7 [00:01<00:03, 1.52it/s][I 2020-09-27 04:43:08,551] Trial 1 finished with value: 0.6512579255125732 and parameters: {'feature_fraction': 1.0}. Best is trial 1 with value: 0.6512579255125732.
feature_fraction, val_score: 0.651258: 29%|##8 | 2/7 [00:01<00:03, 1.52it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004468 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.583108 valid's binary_logloss: 0.653379
Early stopping, best iteration is:
[83] train's binary_logloss: 0.593455 valid's binary_logloss: 0.652673
feature_fraction, val_score: 0.651258: 43%|####2 | 3/7 [00:01<00:02, 1.69it/s][I 2020-09-27 04:43:08,993] Trial 2 finished with value: 0.652673016771765 and parameters: {'feature_fraction': 0.4}. Best is trial 1 with value: 0.6512579255125732.
feature_fraction, val_score: 0.651258: 43%|####2 | 3/7 [00:01<00:02, 1.69it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000894 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.573161 valid's binary_logloss: 0.653887
[200] train's binary_logloss: 0.518397 valid's binary_logloss: 0.656355
Early stopping, best iteration is:
[117] train's binary_logloss: 0.562873 valid's binary_logloss: 0.652985
feature_fraction, val_score: 0.651258: 57%|#####7 | 4/7 [00:02<00:01, 1.66it/s][I 2020-09-27 04:43:09,624] Trial 3 finished with value: 0.6529847756787174 and parameters: {'feature_fraction': 0.8}. Best is trial 1 with value: 0.6512579255125732.
feature_fraction, val_score: 0.651258: 57%|#####7 | 4/7 [00:02<00:01, 1.66it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000779 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.574837 valid's binary_logloss: 0.653256
Early stopping, best iteration is:
[78] train's binary_logloss: 0.589091 valid's binary_logloss: 0.651848
feature_fraction, val_score: 0.651258: 71%|#######1 | 5/7 [00:03<00:01, 1.75it/s][I 2020-09-27 04:43:10,125] Trial 4 finished with value: 0.6518483772036818 and parameters: {'feature_fraction': 0.7}. Best is trial 1 with value: 0.6512579255125732.
feature_fraction, val_score: 0.651258: 71%|#######1 | 5/7 [00:03<00:01, 1.75it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000916 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.572031 valid's binary_logloss: 0.652234
Early stopping, best iteration is:
[76] train's binary_logloss: 0.588368 valid's binary_logloss: 0.651017
feature_fraction, val_score: 0.651017: 86%|########5 | 6/7 [00:03<00:00, 1.79it/s][I 2020-09-27 04:43:10,646] Trial 5 finished with value: 0.6510174434829193 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 5 with value: 0.6510174434829193.
feature_fraction, val_score: 0.651017: 86%|########5 | 6/7 [00:03<00:00, 1.79it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000533 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.578051 valid's binary_logloss: 0.652005
[200] train's binary_logloss: 0.525344 valid's binary_logloss: 0.654096
Early stopping, best iteration is:
[137] train's binary_logloss: 0.557437 valid's binary_logloss: 0.651283
feature_fraction, val_score: 0.651017: 100%|##########| 7/7 [00:04<00:00, 1.76it/s][I 2020-09-27 04:43:11,237] Trial 6 finished with value: 0.6512826266964433 and parameters: {'feature_fraction': 0.5}. Best is trial 5 with value: 0.6510174434829193.
feature_fraction, val_score: 0.651017: 100%|##########| 7/7 [00:04<00:00, 1.65it/s]
num_leaves, val_score: 0.651017: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000858 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.519602 valid's binary_logloss: 0.657482
Early stopping, best iteration is:
[48] train's binary_logloss: 0.578782 valid's binary_logloss: 0.655742
num_leaves, val_score: 0.651017: 5%|5 | 1/20 [00:01<00:23, 1.23s/it][I 2020-09-27 04:43:12,484] Trial 7 finished with value: 0.655741889936831 and parameters: {'num_leaves': 55}. Best is trial 7 with value: 0.655741889936831.
num_leaves, val_score: 0.651017: 5%|5 | 1/20 [00:01<00:23, 1.23s/it][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000474 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.352805 valid's binary_logloss: 0.663625
Early stopping, best iteration is:
[40] train's binary_logloss: 0.492845 valid's binary_logloss: 0.6534
num_leaves, val_score: 0.651017: 10%|# | 2/20 [00:02<00:21, 1.18s/it][I 2020-09-27 04:43:13,542] Trial 8 finished with value: 0.653399633240566 and parameters: {'num_leaves': 164}. Best is trial 8 with value: 0.653399633240566.
num_leaves, val_score: 0.651017: 10%|# | 2/20 [00:02<00:21, 1.18s/it][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005116 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.645865 valid's binary_logloss: 0.655865
[200] train's binary_logloss: 0.633609 valid's binary_logloss: 0.652415
[300] train's binary_logloss: 0.624725 valid's binary_logloss: 0.651323
Early stopping, best iteration is:
[292] train's binary_logloss: 0.625397 valid's binary_logloss: 0.65122
num_leaves, val_score: 0.651017: 15%|#5 | 3/20 [00:02<00:17, 1.02s/it][I 2020-09-27 04:43:14,213] Trial 9 finished with value: 0.6512203104712043 and parameters: {'num_leaves': 5}. Best is trial 9 with value: 0.6512203104712043.
num_leaves, val_score: 0.651017: 15%|#5 | 3/20 [00:02<00:17, 1.02s/it][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000891 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.297872 valid's binary_logloss: 0.671789
Early stopping, best iteration is:
[27] train's binary_logloss: 0.510065 valid's binary_logloss: 0.658243
num_leaves, val_score: 0.651017: 20%|## | 4/20 [00:04<00:20, 1.30s/it][I 2020-09-27 04:43:16,157] Trial 10 finished with value: 0.6582425249764476 and parameters: {'num_leaves': 215}. Best is trial 9 with value: 0.6512203104712043.
num_leaves, val_score: 0.651017: 20%|## | 4/20 [00:04<00:20, 1.30s/it][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000907 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.669067 valid's binary_logloss: 0.671099
[200] train's binary_logloss: 0.659436 valid's binary_logloss: 0.661559
[300] train's binary_logloss: 0.654191 valid's binary_logloss: 0.656836
[400] train's binary_logloss: 0.651007 valid's binary_logloss: 0.65419
[500] train's binary_logloss: 0.64896 valid's binary_logloss: 0.652841
[600] train's binary_logloss: 0.647587 valid's binary_logloss: 0.652048
[700] train's binary_logloss: 0.646612 valid's binary_logloss: 0.651613
[800] train's binary_logloss: 0.645876 valid's binary_logloss: 0.651451
[900] train's binary_logloss: 0.645296 valid's binary_logloss: 0.651421
Early stopping, best iteration is:
[887] train's binary_logloss: 0.645365 valid's binary_logloss: 0.651313
num_leaves, val_score: 0.651017: 25%|##5 | 5/20 [00:06<00:19, 1.29s/it][I 2020-09-27 04:43:17,415] Trial 11 finished with value: 0.6513127631312726 and parameters: {'num_leaves': 2}. Best is trial 9 with value: 0.6512203104712043.
num_leaves, val_score: 0.651017: 25%|##5 | 5/20 [00:06<00:19, 1.29s/it][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001145 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.487358 valid's binary_logloss: 0.659662
Early stopping, best iteration is:
[35] train's binary_logloss: 0.581991 valid's binary_logloss: 0.657768
num_leaves, val_score: 0.651017: 30%|### | 6/20 [00:06<00:15, 1.10s/it][I 2020-09-27 04:43:18,063] Trial 12 finished with value: 0.6577682570812166 and parameters: {'num_leaves': 73}. Best is trial 9 with value: 0.6512203104712043.
num_leaves, val_score: 0.651017: 30%|### | 6/20 [00:06<00:15, 1.10s/it][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001036 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.607358 valid's binary_logloss: 0.650129
[200] train's binary_logloss: 0.572939 valid's binary_logloss: 0.650833
Early stopping, best iteration is:
[137] train's binary_logloss: 0.593957 valid's binary_logloss: 0.649769
num_leaves, val_score: 0.649769: 35%|###5 | 7/20 [00:07<00:12, 1.07it/s][I 2020-09-27 04:43:18,630] Trial 13 finished with value: 0.649768864866138 and parameters: {'num_leaves': 17}. Best is trial 13 with value: 0.649768864866138.
num_leaves, val_score: 0.649769: 35%|###5 | 7/20 [00:07<00:12, 1.07it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001092 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.41967 valid's binary_logloss: 0.663995
Early stopping, best iteration is:
[33] train's binary_logloss: 0.552391 valid's binary_logloss: 0.657367
num_leaves, val_score: 0.649769: 40%|#### | 8/20 [00:08<00:12, 1.07s/it][I 2020-09-27 04:43:20,007] Trial 14 finished with value: 0.6573672765256382 and parameters: {'num_leaves': 114}. Best is trial 13 with value: 0.649768864866138.
num_leaves, val_score: 0.649769: 40%|#### | 8/20 [00:08<00:12, 1.07s/it][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000964 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.262043 valid's binary_logloss: 0.668022
Early stopping, best iteration is:
[31] train's binary_logloss: 0.470728 valid's binary_logloss: 0.656929
num_leaves, val_score: 0.649769: 45%|####5 | 9/20 [00:10<00:13, 1.26s/it][I 2020-09-27 04:43:21,707] Trial 15 finished with value: 0.6569287641382702 and parameters: {'num_leaves': 256}. Best is trial 13 with value: 0.649768864866138.
num_leaves, val_score: 0.649769: 45%|####5 | 9/20 [00:10<00:13, 1.26s/it][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000921 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.534752 valid's binary_logloss: 0.65496
Early stopping, best iteration is:
[63] train's binary_logloss: 0.569855 valid's binary_logloss: 0.652951
num_leaves, val_score: 0.649769: 50%|##### | 10/20 [00:11<00:10, 1.07s/it][I 2020-09-27 04:43:22,329] Trial 16 finished with value: 0.6529514815127009 and parameters: {'num_leaves': 48}. Best is trial 13 with value: 0.649768864866138.
num_leaves, val_score: 0.649769: 50%|##### | 10/20 [00:11<00:10, 1.07s/it][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004780 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.382444 valid's binary_logloss: 0.663277
Early stopping, best iteration is:
[36] train's binary_logloss: 0.523542 valid's binary_logloss: 0.654707
num_leaves, val_score: 0.649769: 55%|#####5 | 11/20 [00:12<00:10, 1.12s/it][I 2020-09-27 04:43:23,589] Trial 17 finished with value: 0.6547072669652428 and parameters: {'num_leaves': 140}. Best is trial 13 with value: 0.649768864866138.
num_leaves, val_score: 0.649769: 55%|#####5 | 11/20 [00:12<00:10, 1.12s/it][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.003413 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.448357 valid's binary_logloss: 0.661818
Early stopping, best iteration is:
[46] train's binary_logloss: 0.538243 valid's binary_logloss: 0.657096
num_leaves, val_score: 0.649769: 60%|###### | 12/20 [00:13<00:08, 1.08s/it][I 2020-09-27 04:43:24,567] Trial 18 finished with value: 0.6570959414584372 and parameters: {'num_leaves': 95}. Best is trial 13 with value: 0.649768864866138.
num_leaves, val_score: 0.649769: 60%|###### | 12/20 [00:13<00:08, 1.08s/it][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010468 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.589787 valid's binary_logloss: 0.652218
Early stopping, best iteration is:
[97] train's binary_logloss: 0.591353 valid's binary_logloss: 0.652064
num_leaves, val_score: 0.649769: 65%|######5 | 13/20 [00:13<00:06, 1.09it/s][I 2020-09-27 04:43:25,099] Trial 19 finished with value: 0.6520641906266819 and parameters: {'num_leaves': 24}. Best is trial 13 with value: 0.649768864866138.
num_leaves, val_score: 0.649769: 65%|######5 | 13/20 [00:13<00:06, 1.09it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000894 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.332931 valid's binary_logloss: 0.669943
Early stopping, best iteration is:
[24] train's binary_logloss: 0.541056 valid's binary_logloss: 0.658153
num_leaves, val_score: 0.649769: 70%|####### | 14/20 [00:14<00:05, 1.02it/s][I 2020-09-27 04:43:26,213] Trial 20 finished with value: 0.6581527972889681 and parameters: {'num_leaves': 181}. Best is trial 13 with value: 0.649768864866138.
num_leaves, val_score: 0.649769: 70%|####### | 14/20 [00:14<00:05, 1.02it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010464 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.645865 valid's binary_logloss: 0.655865
[200] train's binary_logloss: 0.633609 valid's binary_logloss: 0.652415
[300] train's binary_logloss: 0.624725 valid's binary_logloss: 0.651323
Early stopping, best iteration is:
[292] train's binary_logloss: 0.625397 valid's binary_logloss: 0.65122
num_leaves, val_score: 0.649769: 75%|#######5 | 15/20 [00:15<00:04, 1.13it/s][I 2020-09-27 04:43:26,886] Trial 21 finished with value: 0.6512203104712043 and parameters: {'num_leaves': 5}. Best is trial 13 with value: 0.649768864866138.
num_leaves, val_score: 0.649769: 75%|#######5 | 15/20 [00:15<00:04, 1.13it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000872 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.579189 valid's binary_logloss: 0.654451
[200] train's binary_logloss: 0.527779 valid's binary_logloss: 0.653698
Early stopping, best iteration is:
[151] train's binary_logloss: 0.551963 valid's binary_logloss: 0.653462
num_leaves, val_score: 0.649769: 80%|######## | 16/20 [00:16<00:04, 1.00s/it][I 2020-09-27 04:43:28,167] Trial 22 finished with value: 0.6534616704278683 and parameters: {'num_leaves': 28}. Best is trial 13 with value: 0.649768864866138.
num_leaves, val_score: 0.649769: 80%|######## | 16/20 [00:16<00:04, 1.00s/it][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011614 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.472661 valid's binary_logloss: 0.658505
Early stopping, best iteration is:
[28] train's binary_logloss: 0.589272 valid's binary_logloss: 0.658288
num_leaves, val_score: 0.649769: 85%|########5 | 17/20 [00:17<00:02, 1.13it/s][I 2020-09-27 04:43:28,784] Trial 23 finished with value: 0.658288148145066 and parameters: {'num_leaves': 81}. Best is trial 13 with value: 0.649768864866138.
num_leaves, val_score: 0.649769: 85%|########5 | 17/20 [00:17<00:02, 1.13it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000925 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.645865 valid's binary_logloss: 0.655865
[200] train's binary_logloss: 0.633609 valid's binary_logloss: 0.652415
[300] train's binary_logloss: 0.624725 valid's binary_logloss: 0.651323
Early stopping, best iteration is:
[292] train's binary_logloss: 0.625397 valid's binary_logloss: 0.65122
num_leaves, val_score: 0.649769: 90%|######### | 18/20 [00:18<00:01, 1.23it/s][I 2020-09-27 04:43:29,433] Trial 24 finished with value: 0.6512203104712043 and parameters: {'num_leaves': 5}. Best is trial 13 with value: 0.649768864866138.
num_leaves, val_score: 0.649769: 90%|######### | 18/20 [00:18<00:01, 1.23it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004987 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.565987 valid's binary_logloss: 0.655697
Early stopping, best iteration is:
[59] train's binary_logloss: 0.596142 valid's binary_logloss: 0.65418
num_leaves, val_score: 0.649769: 95%|#########5| 19/20 [00:18<00:00, 1.34it/s][I 2020-09-27 04:43:30,009] Trial 25 finished with value: 0.6541801455639907 and parameters: {'num_leaves': 34}. Best is trial 13 with value: 0.649768864866138.
num_leaves, val_score: 0.649769: 95%|#########5| 19/20 [00:18<00:00, 1.34it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010666 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.512496 valid's binary_logloss: 0.657066
Early stopping, best iteration is:
[78] train's binary_logloss: 0.536548 valid's binary_logloss: 0.654935
num_leaves, val_score: 0.649769: 100%|##########| 20/20 [00:19<00:00, 1.38it/s][I 2020-09-27 04:43:30,691] Trial 26 finished with value: 0.6549352289587036 and parameters: {'num_leaves': 59}. Best is trial 13 with value: 0.649768864866138.
num_leaves, val_score: 0.649769: 100%|##########| 20/20 [00:19<00:00, 1.03it/s]
bagging, val_score: 0.649769: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000913 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.612583 valid's binary_logloss: 0.656379
Early stopping, best iteration is:
[86] train's binary_logloss: 0.617678 valid's binary_logloss: 0.655144
bagging, val_score: 0.649769: 10%|# | 1/10 [00:00<00:07, 1.17it/s][I 2020-09-27 04:43:31,563] Trial 27 finished with value: 0.6551437821744738 and parameters: {'bagging_fraction': 0.40320944016041527, 'bagging_freq': 7}. Best is trial 27 with value: 0.6551437821744738.
bagging, val_score: 0.649769: 10%|# | 1/10 [00:00<00:07, 1.17it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.014340 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.607438 valid's binary_logloss: 0.653472
[200] train's binary_logloss: 0.573071 valid's binary_logloss: 0.653346
Early stopping, best iteration is:
[139] train's binary_logloss: 0.593171 valid's binary_logloss: 0.652081
bagging, val_score: 0.649769: 20%|## | 2/10 [00:01<00:06, 1.21it/s][I 2020-09-27 04:43:32,320] Trial 28 finished with value: 0.6520811290472097 and parameters: {'bagging_fraction': 0.9561159020668242, 'bagging_freq': 1}. Best is trial 28 with value: 0.6520811290472097.
bagging, val_score: 0.649769: 20%|## | 2/10 [00:01<00:06, 1.21it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001122 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.610187 valid's binary_logloss: 0.652923
Early stopping, best iteration is:
[81] train's binary_logloss: 0.617563 valid's binary_logloss: 0.651804
bagging, val_score: 0.649769: 30%|### | 3/10 [00:02<00:05, 1.39it/s][I 2020-09-27 04:43:32,796] Trial 29 finished with value: 0.6518044921241909 and parameters: {'bagging_fraction': 0.5465375900437621, 'bagging_freq': 4}. Best is trial 29 with value: 0.6518044921241909.
bagging, val_score: 0.649769: 30%|### | 3/10 [00:02<00:05, 1.39it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002328 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.607479 valid's binary_logloss: 0.651182
[200] train's binary_logloss: 0.571824 valid's binary_logloss: 0.653082
Early stopping, best iteration is:
[118] train's binary_logloss: 0.600602 valid's binary_logloss: 0.650679
bagging, val_score: 0.649769: 40%|#### | 4/10 [00:02<00:04, 1.46it/s][I 2020-09-27 04:43:33,403] Trial 30 finished with value: 0.6506791175850941 and parameters: {'bagging_fraction': 0.9711225430872865, 'bagging_freq': 7}. Best is trial 30 with value: 0.6506791175850941.
bagging, val_score: 0.649769: 40%|#### | 4/10 [00:02<00:04, 1.46it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005891 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.607508 valid's binary_logloss: 0.652005
[200] train's binary_logloss: 0.572706 valid's binary_logloss: 0.650621
Early stopping, best iteration is:
[129] train's binary_logloss: 0.596726 valid's binary_logloss: 0.650382
bagging, val_score: 0.649769: 50%|##### | 5/10 [00:03<00:03, 1.48it/s][I 2020-09-27 04:43:34,052] Trial 31 finished with value: 0.6503824259794613 and parameters: {'bagging_fraction': 0.9985128973254092, 'bagging_freq': 7}. Best is trial 31 with value: 0.6503824259794613.
bagging, val_score: 0.649769: 50%|##### | 5/10 [00:03<00:03, 1.48it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005483 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.60784 valid's binary_logloss: 0.652715
[200] train's binary_logloss: 0.572748 valid's binary_logloss: 0.653478
Early stopping, best iteration is:
[124] train's binary_logloss: 0.598892 valid's binary_logloss: 0.651677
bagging, val_score: 0.649769: 60%|###### | 6/10 [00:04<00:02, 1.48it/s][I 2020-09-27 04:43:34,731] Trial 32 finished with value: 0.6516771078927124 and parameters: {'bagging_fraction': 0.9937517875557063, 'bagging_freq': 7}. Best is trial 31 with value: 0.6503824259794613.
bagging, val_score: 0.649769: 60%|###### | 6/10 [00:04<00:02, 1.48it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.003559 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.607577 valid's binary_logloss: 0.652062
[200] train's binary_logloss: 0.572523 valid's binary_logloss: 0.651285
[300] train's binary_logloss: 0.541985 valid's binary_logloss: 0.651463
Early stopping, best iteration is:
[260] train's binary_logloss: 0.554233 valid's binary_logloss: 0.65052
bagging, val_score: 0.649769: 70%|####### | 7/10 [00:05<00:02, 1.13it/s][I 2020-09-27 04:43:36,093] Trial 33 finished with value: 0.6505203576283998 and parameters: {'bagging_fraction': 0.8468420031688184, 'bagging_freq': 6}. Best is trial 31 with value: 0.6503824259794613.
bagging, val_score: 0.649769: 70%|####### | 7/10 [00:05<00:02, 1.13it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001043 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.60809 valid's binary_logloss: 0.653721
Early stopping, best iteration is:
[83] train's binary_logloss: 0.614817 valid's binary_logloss: 0.653287
bagging, val_score: 0.649769: 80%|######## | 8/10 [00:05<00:01, 1.30it/s][I 2020-09-27 04:43:36,593] Trial 34 finished with value: 0.6532867425106659 and parameters: {'bagging_fraction': 0.8464729466696903, 'bagging_freq': 7}. Best is trial 31 with value: 0.6503824259794613.
bagging, val_score: 0.649769: 80%|######## | 8/10 [00:05<00:01, 1.30it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000488 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.608096 valid's binary_logloss: 0.652476
[200] train's binary_logloss: 0.572069 valid's binary_logloss: 0.654848
Early stopping, best iteration is:
[105] train's binary_logloss: 0.606068 valid's binary_logloss: 0.652367
bagging, val_score: 0.649769: 90%|######### | 9/10 [00:06<00:00, 1.41it/s][I 2020-09-27 04:43:37,161] Trial 35 finished with value: 0.6523668340051799 and parameters: {'bagging_fraction': 0.830591192364971, 'bagging_freq': 5}. Best is trial 31 with value: 0.6503824259794613.
bagging, val_score: 0.649769: 90%|######### | 9/10 [00:06<00:00, 1.41it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000998 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.607246 valid's binary_logloss: 0.652045
Early stopping, best iteration is:
[95] train's binary_logloss: 0.609255 valid's binary_logloss: 0.651848
bagging, val_score: 0.649769: 100%|##########| 10/10 [00:07<00:00, 1.50it/s][I 2020-09-27 04:43:37,726] Trial 36 finished with value: 0.6518478119160271 and parameters: {'bagging_fraction': 0.9992112947235333, 'bagging_freq': 6}. Best is trial 31 with value: 0.6503824259794613.
bagging, val_score: 0.649769: 100%|##########| 10/10 [00:07<00:00, 1.42it/s]
feature_fraction_stage2, val_score: 0.649769: 0%| | 0/6 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000951 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.606705 valid's binary_logloss: 0.652233
[200] train's binary_logloss: 0.572023 valid's binary_logloss: 0.650662
[300] train's binary_logloss: 0.542509 valid's binary_logloss: 0.653346
Early stopping, best iteration is:
[200] train's binary_logloss: 0.572023 valid's binary_logloss: 0.650662
feature_fraction_stage2, val_score: 0.649769: 17%|#6 | 1/6 [00:00<00:03, 1.42it/s][I 2020-09-27 04:43:38,443] Trial 37 finished with value: 0.6506622068727096 and parameters: {'feature_fraction': 0.9799999999999999}. Best is trial 37 with value: 0.6506622068727096.
feature_fraction_stage2, val_score: 0.649769: 17%|#6 | 1/6 [00:00<00:03, 1.42it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000943 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.607972 valid's binary_logloss: 0.651443
[200] train's binary_logloss: 0.573729 valid's binary_logloss: 0.651681
Early stopping, best iteration is:
[174] train's binary_logloss: 0.58195 valid's binary_logloss: 0.650991
feature_fraction_stage2, val_score: 0.649769: 33%|###3 | 2/6 [00:01<00:03, 1.17it/s][I 2020-09-27 04:43:39,648] Trial 38 finished with value: 0.6509908340902059 and parameters: {'feature_fraction': 0.852}. Best is trial 37 with value: 0.6506622068727096.
feature_fraction_stage2, val_score: 0.649769: 33%|###3 | 2/6 [00:01<00:03, 1.17it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000929 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.606705 valid's binary_logloss: 0.652233
[200] train's binary_logloss: 0.572023 valid's binary_logloss: 0.650662
[300] train's binary_logloss: 0.542509 valid's binary_logloss: 0.653346
Early stopping, best iteration is:
[200] train's binary_logloss: 0.572023 valid's binary_logloss: 0.650662
feature_fraction_stage2, val_score: 0.649769: 50%|##### | 3/6 [00:02<00:02, 1.22it/s][I 2020-09-27 04:43:40,396] Trial 39 finished with value: 0.6506622068727096 and parameters: {'feature_fraction': 0.948}. Best is trial 37 with value: 0.6506622068727096.
feature_fraction_stage2, val_score: 0.649769: 50%|##### | 3/6 [00:02<00:02, 1.22it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000871 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.60867 valid's binary_logloss: 0.651571
Early stopping, best iteration is:
[95] train's binary_logloss: 0.610608 valid's binary_logloss: 0.651188
feature_fraction_stage2, val_score: 0.649769: 67%|######6 | 4/6 [00:03<00:01, 1.37it/s][I 2020-09-27 04:43:40,904] Trial 40 finished with value: 0.6511879268615787 and parameters: {'feature_fraction': 0.82}. Best is trial 37 with value: 0.6506622068727096.
feature_fraction_stage2, val_score: 0.649769: 67%|######6 | 4/6 [00:03<00:01, 1.37it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005317 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.60735 valid's binary_logloss: 0.652555
Early stopping, best iteration is:
[72] train's binary_logloss: 0.618944 valid's binary_logloss: 0.651968
feature_fraction_stage2, val_score: 0.649769: 83%|########3 | 5/6 [00:03<00:00, 1.53it/s][I 2020-09-27 04:43:41,380] Trial 41 finished with value: 0.6519681937770695 and parameters: {'feature_fraction': 0.9159999999999999}. Best is trial 37 with value: 0.6506622068727096.
feature_fraction_stage2, val_score: 0.649769: 83%|########3 | 5/6 [00:03<00:00, 1.53it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005114 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.607358 valid's binary_logloss: 0.650129
[200] train's binary_logloss: 0.572939 valid's binary_logloss: 0.650833
Early stopping, best iteration is:
[137] train's binary_logloss: 0.593957 valid's binary_logloss: 0.649769
feature_fraction_stage2, val_score: 0.649769: 100%|##########| 6/6 [00:04<00:00, 1.54it/s][I 2020-09-27 04:43:42,023] Trial 42 finished with value: 0.649768864866138 and parameters: {'feature_fraction': 0.8839999999999999}. Best is trial 42 with value: 0.649768864866138.
feature_fraction_stage2, val_score: 0.649769: 100%|##########| 6/6 [00:04<00:00, 1.40it/s]
regularization_factors, val_score: 0.649769: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000939 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.607973 valid's binary_logloss: 0.65115
[200] train's binary_logloss: 0.573107 valid's binary_logloss: 0.649872
[300] train's binary_logloss: 0.543329 valid's binary_logloss: 0.651302
Early stopping, best iteration is:
[229] train's binary_logloss: 0.564107 valid's binary_logloss: 0.649364
regularization_factors, val_score: 0.649364: 5%|5 | 1/20 [00:01<00:24, 1.27s/it][I 2020-09-27 04:43:43,315] Trial 43 finished with value: 0.6493642189216182 and parameters: {'lambda_l1': 0.16783910794593826, 'lambda_l2': 6.907642804430548e-07}. Best is trial 43 with value: 0.6493642189216182.
regularization_factors, val_score: 0.649364: 5%|5 | 1/20 [00:01<00:24, 1.27s/it][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000994 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.608181 valid's binary_logloss: 0.65244
[200] train's binary_logloss: 0.574199 valid's binary_logloss: 0.652483
Early stopping, best iteration is:
[119] train's binary_logloss: 0.601088 valid's binary_logloss: 0.651426
regularization_factors, val_score: 0.649364: 10%|# | 2/20 [00:01<00:19, 1.08s/it][I 2020-09-27 04:43:43,949] Trial 44 finished with value: 0.6514259112904369 and parameters: {'lambda_l1': 0.3129099140173397, 'lambda_l2': 3.3514640861329857e-07}. Best is trial 43 with value: 0.6493642189216182.
regularization_factors, val_score: 0.649364: 10%|# | 2/20 [00:01<00:19, 1.08s/it][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009984 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.607688 valid's binary_logloss: 0.651501
[200] train's binary_logloss: 0.57304 valid's binary_logloss: 0.648731
[300] train's binary_logloss: 0.543627 valid's binary_logloss: 0.650347
Early stopping, best iteration is:
[212] train's binary_logloss: 0.569524 valid's binary_logloss: 0.64867
regularization_factors, val_score: 0.648670: 15%|#5 | 3/20 [00:02<00:16, 1.01it/s][I 2020-09-27 04:43:44,720] Trial 45 finished with value: 0.6486701923102753 and parameters: {'lambda_l1': 7.774361825486131e-06, 'lambda_l2': 0.006013598728332988}. Best is trial 45 with value: 0.6486701923102753.
regularization_factors, val_score: 0.648670: 15%|#5 | 3/20 [00:02<00:16, 1.01it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000855 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.607956 valid's binary_logloss: 0.650205
[200] train's binary_logloss: 0.573144 valid's binary_logloss: 0.651349
Early stopping, best iteration is:
[109] train's binary_logloss: 0.604467 valid's binary_logloss: 0.649709
regularization_factors, val_score: 0.648670: 20%|## | 4/20 [00:03<00:13, 1.17it/s][I 2020-09-27 04:43:45,259] Trial 46 finished with value: 0.6497086221332595 and parameters: {'lambda_l1': 4.0595395206550105e-07, 'lambda_l2': 0.16859061913783596}. Best is trial 45 with value: 0.6486701923102753.
regularization_factors, val_score: 0.648670: 20%|## | 4/20 [00:03<00:13, 1.17it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000898 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.607896 valid's binary_logloss: 0.649423
[200] train's binary_logloss: 0.574175 valid's binary_logloss: 0.64969
Early stopping, best iteration is:
[121] train's binary_logloss: 0.600172 valid's binary_logloss: 0.648913
regularization_factors, val_score: 0.648670: 25%|##5 | 5/20 [00:03<00:11, 1.29it/s][I 2020-09-27 04:43:45,841] Trial 47 finished with value: 0.6489130270973635 and parameters: {'lambda_l1': 1.9089796516560767e-07, 'lambda_l2': 0.2199091551428564}. Best is trial 45 with value: 0.6486701923102753.
regularization_factors, val_score: 0.648670: 25%|##5 | 5/20 [00:03<00:11, 1.29it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000906 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.607889 valid's binary_logloss: 0.650079
[200] train's binary_logloss: 0.574212 valid's binary_logloss: 0.65031
Early stopping, best iteration is:
[131] train's binary_logloss: 0.59635 valid's binary_logloss: 0.648762
regularization_factors, val_score: 0.648670: 30%|### | 6/20 [00:04<00:11, 1.23it/s][I 2020-09-27 04:43:46,751] Trial 48 finished with value: 0.6487624267173515 and parameters: {'lambda_l1': 4.2319368851060655e-07, 'lambda_l2': 0.34004310864421733}. Best is trial 45 with value: 0.6486701923102753.
regularization_factors, val_score: 0.648670: 30%|### | 6/20 [00:04<00:11, 1.23it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.014681 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.608275 valid's binary_logloss: 0.650944
[200] train's binary_logloss: 0.573416 valid's binary_logloss: 0.651579
Early stopping, best iteration is:
[113] train's binary_logloss: 0.603189 valid's binary_logloss: 0.650489
regularization_factors, val_score: 0.648670: 35%|###5 | 7/20 [00:05<00:10, 1.22it/s][I 2020-09-27 04:43:47,591] Trial 49 finished with value: 0.6504888821768885 and parameters: {'lambda_l1': 1.875716331797824e-07, 'lambda_l2': 0.3304208617133868}. Best is trial 45 with value: 0.6486701923102753.
regularization_factors, val_score: 0.648670: 35%|###5 | 7/20 [00:05<00:10, 1.22it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001188 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.608185 valid's binary_logloss: 0.650237
Early stopping, best iteration is:
[98] train's binary_logloss: 0.608923 valid's binary_logloss: 0.650108
regularization_factors, val_score: 0.648670: 40%|#### | 8/20 [00:06<00:08, 1.35it/s][I 2020-09-27 04:43:48,143] Trial 50 finished with value: 0.6501084626471252 and parameters: {'lambda_l1': 4.671913690463379e-07, 'lambda_l2': 0.1995460595674108}. Best is trial 45 with value: 0.6486701923102753.
regularization_factors, val_score: 0.648670: 40%|#### | 8/20 [00:06<00:08, 1.35it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000823 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.60758 valid's binary_logloss: 0.651374
[200] train's binary_logloss: 0.57306 valid's binary_logloss: 0.651712
Early stopping, best iteration is:
[154] train's binary_logloss: 0.588205 valid's binary_logloss: 0.650518
regularization_factors, val_score: 0.648670: 45%|####5 | 9/20 [00:06<00:07, 1.39it/s][I 2020-09-27 04:43:48,805] Trial 51 finished with value: 0.6505179255791086 and parameters: {'lambda_l1': 7.519476866197216e-07, 'lambda_l2': 0.08987068172093761}. Best is trial 45 with value: 0.6486701923102753.
regularization_factors, val_score: 0.648670: 45%|####5 | 9/20 [00:06<00:07, 1.39it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000917 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.607423 valid's binary_logloss: 0.652085
[200] train's binary_logloss: 0.573345 valid's binary_logloss: 0.651502
Early stopping, best iteration is:
[183] train's binary_logloss: 0.578885 valid's binary_logloss: 0.651003
regularization_factors, val_score: 0.648670: 50%|##### | 10/20 [00:07<00:07, 1.40it/s][I 2020-09-27 04:43:49,511] Trial 52 finished with value: 0.6510031934822363 and parameters: {'lambda_l1': 7.307330460252661e-07, 'lambda_l2': 0.0186645031244208}. Best is trial 45 with value: 0.6486701923102753.
regularization_factors, val_score: 0.648670: 50%|##### | 10/20 [00:07<00:07, 1.40it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002545 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.607358 valid's binary_logloss: 0.650129
[200] train's binary_logloss: 0.572941 valid's binary_logloss: 0.650864
Early stopping, best iteration is:
[137] train's binary_logloss: 0.593957 valid's binary_logloss: 0.649769
regularization_factors, val_score: 0.648670: 55%|#####5 | 11/20 [00:08<00:06, 1.46it/s][I 2020-09-27 04:43:50,135] Trial 53 finished with value: 0.6497687989983096 and parameters: {'lambda_l1': 0.0002292384736520708, 'lambda_l2': 8.689375650085261e-05}. Best is trial 45 with value: 0.6486701923102753.
regularization_factors, val_score: 0.648670: 55%|#####5 | 11/20 [00:08<00:06, 1.46it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000469 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.607679 valid's binary_logloss: 0.651246
[200] train's binary_logloss: 0.57339 valid's binary_logloss: 0.653045
Early stopping, best iteration is:
[127] train's binary_logloss: 0.597662 valid's binary_logloss: 0.650697
regularization_factors, val_score: 0.648670: 60%|###### | 12/20 [00:09<00:06, 1.23it/s][I 2020-09-27 04:43:51,237] Trial 54 finished with value: 0.6506972734005597 and parameters: {'lambda_l1': 0.001430301077955183, 'lambda_l2': 5.329476633929677e-06}. Best is trial 45 with value: 0.6486701923102753.
regularization_factors, val_score: 0.648670: 60%|###### | 12/20 [00:09<00:06, 1.23it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000893 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.607359 valid's binary_logloss: 0.650129
[200] train's binary_logloss: 0.572942 valid's binary_logloss: 0.650833
Early stopping, best iteration is:
[137] train's binary_logloss: 0.593959 valid's binary_logloss: 0.649769
regularization_factors, val_score: 0.648670: 65%|######5 | 13/20 [00:09<00:05, 1.31it/s][I 2020-09-27 04:43:51,881] Trial 55 finished with value: 0.6497686874558546 and parameters: {'lambda_l1': 0.00011580371818856396, 'lambda_l2': 0.0005902239811498276}. Best is trial 45 with value: 0.6486701923102753.
regularization_factors, val_score: 0.648670: 65%|######5 | 13/20 [00:09<00:05, 1.31it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.007667 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.607682 valid's binary_logloss: 0.651246
[200] train's binary_logloss: 0.57384 valid's binary_logloss: 0.650954
[300] train's binary_logloss: 0.543099 valid's binary_logloss: 0.650065
Early stopping, best iteration is:
[273] train's binary_logloss: 0.551078 valid's binary_logloss: 0.649674
regularization_factors, val_score: 0.648670: 70%|####### | 14/20 [00:10<00:04, 1.25it/s][I 2020-09-27 04:43:52,762] Trial 56 finished with value: 0.649674471194536 and parameters: {'lambda_l1': 3.268557145355658e-05, 'lambda_l2': 0.003069238534147287}. Best is trial 45 with value: 0.6486701923102753.
regularization_factors, val_score: 0.648670: 70%|####### | 14/20 [00:10<00:04, 1.25it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000897 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.612768 valid's binary_logloss: 0.651575
[200] train's binary_logloss: 0.586233 valid's binary_logloss: 0.651305
Early stopping, best iteration is:
[160] train's binary_logloss: 0.595921 valid's binary_logloss: 0.650515
regularization_factors, val_score: 0.648670: 75%|#######5 | 15/20 [00:11<00:03, 1.31it/s][I 2020-09-27 04:43:53,439] Trial 57 finished with value: 0.6505145838820116 and parameters: {'lambda_l1': 5.615730195745978e-06, 'lambda_l2': 7.365468833430753}. Best is trial 45 with value: 0.6486701923102753.
regularization_factors, val_score: 0.648670: 75%|#######5 | 15/20 [00:11<00:03, 1.31it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.007481 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.60768 valid's binary_logloss: 0.651246
[200] train's binary_logloss: 0.573394 valid's binary_logloss: 0.653045
Early stopping, best iteration is:
[127] train's binary_logloss: 0.597664 valid's binary_logloss: 0.650697
regularization_factors, val_score: 0.648670: 80%|######## | 16/20 [00:12<00:02, 1.38it/s][I 2020-09-27 04:43:54,076] Trial 58 finished with value: 0.6506971138019576 and parameters: {'lambda_l1': 2.3099984324631693e-08, 'lambda_l2': 0.0023773905893283293}. Best is trial 45 with value: 0.6486701923102753.
regularization_factors, val_score: 0.648670: 80%|######## | 16/20 [00:12<00:02, 1.38it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001747 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.61046 valid's binary_logloss: 0.65107
[200] train's binary_logloss: 0.581203 valid's binary_logloss: 0.651453
Early stopping, best iteration is:
[171] train's binary_logloss: 0.588919 valid's binary_logloss: 0.650336
regularization_factors, val_score: 0.648670: 85%|########5 | 17/20 [00:13<00:02, 1.15it/s][I 2020-09-27 04:43:55,278] Trial 59 finished with value: 0.6503360249834589 and parameters: {'lambda_l1': 1.2365048737402212e-05, 'lambda_l2': 3.1664647322579405}. Best is trial 45 with value: 0.6486701923102753.
regularization_factors, val_score: 0.648670: 85%|########5 | 17/20 [00:13<00:02, 1.15it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000874 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.607694 valid's binary_logloss: 0.6515
[200] train's binary_logloss: 0.573475 valid's binary_logloss: 0.651154
Early stopping, best iteration is:
[133] train's binary_logloss: 0.595863 valid's binary_logloss: 0.650792
regularization_factors, val_score: 0.648670: 90%|######### | 18/20 [00:13<00:01, 1.27it/s][I 2020-09-27 04:43:55,883] Trial 60 finished with value: 0.6507917448366202 and parameters: {'lambda_l1': 1.4432590771728913e-08, 'lambda_l2': 0.01029217837602668}. Best is trial 45 with value: 0.6486701923102753.
regularization_factors, val_score: 0.648670: 90%|######### | 18/20 [00:13<00:01, 1.27it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000902 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.607359 valid's binary_logloss: 0.650129
[200] train's binary_logloss: 0.572942 valid's binary_logloss: 0.650864
Early stopping, best iteration is:
[137] train's binary_logloss: 0.593958 valid's binary_logloss: 0.649769
regularization_factors, val_score: 0.648670: 95%|#########5| 19/20 [00:14<00:00, 1.34it/s][I 2020-09-27 04:43:56,523] Trial 61 finished with value: 0.6497687138706673 and parameters: {'lambda_l1': 0.00010150982895717758, 'lambda_l2': 0.0005002352741242428}. Best is trial 45 with value: 0.6486701923102753.
regularization_factors, val_score: 0.648670: 95%|#########5| 19/20 [00:14<00:00, 1.34it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000964 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.607358 valid's binary_logloss: 0.650129
[200] train's binary_logloss: 0.57294 valid's binary_logloss: 0.650849
Early stopping, best iteration is:
[137] train's binary_logloss: 0.593957 valid's binary_logloss: 0.649769
regularization_factors, val_score: 0.648670: 100%|##########| 20/20 [00:15<00:00, 1.42it/s][I 2020-09-27 04:43:57,130] Trial 62 finished with value: 0.6497688040658035 and parameters: {'lambda_l1': 1.4113098720716957e-05, 'lambda_l2': 0.00022029341576689507}. Best is trial 45 with value: 0.6486701923102753.
regularization_factors, val_score: 0.648670: 100%|##########| 20/20 [00:15<00:00, 1.32it/s]
min_data_in_leaf, val_score: 0.648670: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000908 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.607864 valid's binary_logloss: 0.650842
[200] train's binary_logloss: 0.573843 valid's binary_logloss: 0.649839
[300] train's binary_logloss: 0.544181 valid's binary_logloss: 0.65183
Early stopping, best iteration is:
[206] train's binary_logloss: 0.572189 valid's binary_logloss: 0.649438
min_data_in_leaf, val_score: 0.648670: 20%|## | 1/5 [00:00<00:03, 1.26it/s][I 2020-09-27 04:43:57,935] Trial 63 finished with value: 0.6494379643673076 and parameters: {'min_child_samples': 25}. Best is trial 63 with value: 0.6494379643673076.
min_data_in_leaf, val_score: 0.648670: 20%|## | 1/5 [00:00<00:03, 1.26it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012759 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.610343 valid's binary_logloss: 0.652543
Early stopping, best iteration is:
[69] train's binary_logloss: 0.622017 valid's binary_logloss: 0.652141
min_data_in_leaf, val_score: 0.648670: 40%|#### | 2/5 [00:01<00:02, 1.13it/s][I 2020-09-27 04:43:59,052] Trial 64 finished with value: 0.6521414167861022 and parameters: {'min_child_samples': 100}. Best is trial 63 with value: 0.6494379643673076.
min_data_in_leaf, val_score: 0.648670: 40%|#### | 2/5 [00:01<00:02, 1.13it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000904 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.60911 valid's binary_logloss: 0.649506
[200] train's binary_logloss: 0.576093 valid's binary_logloss: 0.651061
Early stopping, best iteration is:
[123] train's binary_logloss: 0.601092 valid's binary_logloss: 0.648901
min_data_in_leaf, val_score: 0.648670: 60%|###### | 3/5 [00:02<00:01, 1.23it/s][I 2020-09-27 04:43:59,681] Trial 65 finished with value: 0.6489009066838253 and parameters: {'min_child_samples': 50}. Best is trial 65 with value: 0.6489009066838253.
min_data_in_leaf, val_score: 0.648670: 60%|###### | 3/5 [00:02<00:01, 1.23it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000900 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.607172 valid's binary_logloss: 0.65
[200] train's binary_logloss: 0.572515 valid's binary_logloss: 0.651594
Early stopping, best iteration is:
[100] train's binary_logloss: 0.607172 valid's binary_logloss: 0.65
min_data_in_leaf, val_score: 0.648670: 80%|######## | 4/5 [00:03<00:00, 1.33it/s][I 2020-09-27 04:44:00,293] Trial 66 finished with value: 0.6500004234119392 and parameters: {'min_child_samples': 10}. Best is trial 65 with value: 0.6489009066838253.
min_data_in_leaf, val_score: 0.648670: 80%|######## | 4/5 [00:03<00:00, 1.33it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000902 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4239
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695
[LightGBM] [Info] Start training from score -0.028695
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.607283 valid's binary_logloss: 0.652029
[200] train's binary_logloss: 0.572898 valid's binary_logloss: 0.653554
Early stopping, best iteration is:
[137] train's binary_logloss: 0.593598 valid's binary_logloss: 0.650833
min_data_in_leaf, val_score: 0.648670: 100%|##########| 5/5 [00:03<00:00, 1.38it/s][I 2020-09-27 04:44:00,964] Trial 67 finished with value: 0.6508328689277415 and parameters: {'min_child_samples': 5}. Best is trial 65 with value: 0.6489009066838253.
min_data_in_leaf, val_score: 0.648670: 100%|##########| 5/5 [00:03<00:00, 1.31it/s]
Fold : 6
[I 2020-09-27 04:44:01,016] A new study created in memory with name: no-name-d7eb95f7-845c-40a3-9583-af27505411ce
feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.003088 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.574382 valid's binary_logloss: 0.658832
Early stopping, best iteration is:
[54] train's binary_logloss: 0.607296 valid's binary_logloss: 0.657769
feature_fraction, val_score: 0.657769: 14%|#4 | 1/7 [00:00<00:03, 1.69it/s][I 2020-09-27 04:44:01,616] Trial 0 finished with value: 0.6577689712084953 and parameters: {'feature_fraction': 0.7}. Best is trial 0 with value: 0.6577689712084953.
feature_fraction, val_score: 0.657769: 14%|#4 | 1/7 [00:00<00:03, 1.69it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004999 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.573426 valid's binary_logloss: 0.659195
Early stopping, best iteration is:
[52] train's binary_logloss: 0.607667 valid's binary_logloss: 0.65734
feature_fraction, val_score: 0.657340: 29%|##8 | 2/7 [00:01<00:03, 1.50it/s][I 2020-09-27 04:44:02,465] Trial 1 finished with value: 0.6573399663499826 and parameters: {'feature_fraction': 0.8}. Best is trial 1 with value: 0.6573399663499826.
feature_fraction, val_score: 0.657340: 29%|##8 | 2/7 [00:01<00:03, 1.50it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012514 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.578959 valid's binary_logloss: 0.658522
Early stopping, best iteration is:
[92] train's binary_logloss: 0.583977 valid's binary_logloss: 0.657645
feature_fraction, val_score: 0.657340: 43%|####2 | 3/7 [00:02<00:02, 1.48it/s][I 2020-09-27 04:44:03,161] Trial 2 finished with value: 0.6576453641626878 and parameters: {'feature_fraction': 0.5}. Best is trial 1 with value: 0.6573399663499826.
feature_fraction, val_score: 0.657340: 43%|####2 | 3/7 [00:02<00:02, 1.48it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000380 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.582749 valid's binary_logloss: 0.655485
[200] train's binary_logloss: 0.531378 valid's binary_logloss: 0.660147
Early stopping, best iteration is:
[107] train's binary_logloss: 0.57849 valid's binary_logloss: 0.654569
feature_fraction, val_score: 0.654569: 57%|#####7 | 4/7 [00:03<00:02, 1.04it/s][I 2020-09-27 04:44:04,801] Trial 3 finished with value: 0.654569007042944 and parameters: {'feature_fraction': 0.4}. Best is trial 3 with value: 0.654569007042944.
feature_fraction, val_score: 0.654569: 57%|#####7 | 4/7 [00:03<00:02, 1.04it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000972 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.572042 valid's binary_logloss: 0.658846
Early stopping, best iteration is:
[79] train's binary_logloss: 0.586097 valid's binary_logloss: 0.65804
feature_fraction, val_score: 0.654569: 71%|#######1 | 5/7 [00:04<00:01, 1.16it/s][I 2020-09-27 04:44:05,420] Trial 4 finished with value: 0.6580401508293939 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 3 with value: 0.654569007042944.
feature_fraction, val_score: 0.654569: 71%|#######1 | 5/7 [00:04<00:01, 1.16it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000929 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.571857 valid's binary_logloss: 0.659029
Early stopping, best iteration is:
[44] train's binary_logloss: 0.612853 valid's binary_logloss: 0.657339
feature_fraction, val_score: 0.654569: 86%|########5 | 6/7 [00:04<00:00, 1.30it/s][I 2020-09-27 04:44:05,979] Trial 5 finished with value: 0.6573389310156069 and parameters: {'feature_fraction': 1.0}. Best is trial 3 with value: 0.654569007042944.
feature_fraction, val_score: 0.654569: 86%|########5 | 6/7 [00:04<00:00, 1.30it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016710 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.577175 valid's binary_logloss: 0.657844
Early stopping, best iteration is:
[54] train's binary_logloss: 0.608418 valid's binary_logloss: 0.656152
feature_fraction, val_score: 0.654569: 100%|##########| 7/7 [00:05<00:00, 1.24it/s][I 2020-09-27 04:44:06,878] Trial 6 finished with value: 0.6561524021651702 and parameters: {'feature_fraction': 0.6}. Best is trial 3 with value: 0.654569007042944.
feature_fraction, val_score: 0.654569: 100%|##########| 7/7 [00:05<00:00, 1.19it/s]
num_leaves, val_score: 0.654569: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004746 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.384415 valid's binary_logloss: 0.6703
Early stopping, best iteration is:
[44] train's binary_logloss: 0.508564 valid's binary_logloss: 0.662989
num_leaves, val_score: 0.654569: 5%|5 | 1/20 [00:00<00:14, 1.28it/s][I 2020-09-27 04:44:07,670] Trial 7 finished with value: 0.6629889780799463 and parameters: {'num_leaves': 164}. Best is trial 7 with value: 0.6629889780799463.
num_leaves, val_score: 0.654569: 5%|5 | 1/20 [00:00<00:14, 1.28it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000332 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.410682 valid's binary_logloss: 0.66176
Early stopping, best iteration is:
[62] train's binary_logloss: 0.482134 valid's binary_logloss: 0.658251
num_leaves, val_score: 0.654569: 10%|# | 2/20 [00:01<00:14, 1.22it/s][I 2020-09-27 04:44:08,569] Trial 8 finished with value: 0.6582514287091177 and parameters: {'num_leaves': 142}. Best is trial 8 with value: 0.6582514287091177.
num_leaves, val_score: 0.654569: 10%|# | 2/20 [00:01<00:14, 1.22it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000603 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.412047 valid's binary_logloss: 0.659767
Early stopping, best iteration is:
[58] train's binary_logloss: 0.492004 valid's binary_logloss: 0.656213
num_leaves, val_score: 0.654569: 15%|#5 | 3/20 [00:02<00:14, 1.20it/s][I 2020-09-27 04:44:09,432] Trial 9 finished with value: 0.6562133173559287 and parameters: {'num_leaves': 140}. Best is trial 9 with value: 0.6562133173559287.
num_leaves, val_score: 0.654569: 15%|#5 | 3/20 [00:02<00:14, 1.20it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004277 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.608851 valid's binary_logloss: 0.656989
Early stopping, best iteration is:
[87] train's binary_logloss: 0.614406 valid's binary_logloss: 0.656446
num_leaves, val_score: 0.654569: 20%|## | 4/20 [00:03<00:11, 1.37it/s][I 2020-09-27 04:44:09,931] Trial 10 finished with value: 0.6564457093023696 and parameters: {'num_leaves': 19}. Best is trial 9 with value: 0.6562133173559287.
num_leaves, val_score: 0.654569: 20%|## | 4/20 [00:03<00:11, 1.37it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004598 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.301686 valid's binary_logloss: 0.673949
Early stopping, best iteration is:
[31] train's binary_logloss: 0.504321 valid's binary_logloss: 0.666042
num_leaves, val_score: 0.654569: 25%|##5 | 5/20 [00:04<00:13, 1.09it/s][I 2020-09-27 04:44:11,277] Trial 11 finished with value: 0.66604185972251 and parameters: {'num_leaves': 254}. Best is trial 9 with value: 0.6562133173559287.
num_leaves, val_score: 0.654569: 25%|##5 | 5/20 [00:04<00:13, 1.09it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000377 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.618907 valid's binary_logloss: 0.654583
[200] train's binary_logloss: 0.590148 valid's binary_logloss: 0.653268
Early stopping, best iteration is:
[171] train's binary_logloss: 0.597901 valid's binary_logloss: 0.652352
num_leaves, val_score: 0.652352: 30%|### | 6/20 [00:04<00:11, 1.24it/s][I 2020-09-27 04:44:11,818] Trial 12 finished with value: 0.6523523176355726 and parameters: {'num_leaves': 15}. Best is trial 12 with value: 0.6523523176355726.
num_leaves, val_score: 0.652352: 30%|### | 6/20 [00:04<00:11, 1.24it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000404 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.658358 valid's binary_logloss: 0.663406
[200] train's binary_logloss: 0.647972 valid's binary_logloss: 0.655898
[300] train's binary_logloss: 0.642797 valid's binary_logloss: 0.652806
[400] train's binary_logloss: 0.639134 valid's binary_logloss: 0.651593
[500] train's binary_logloss: 0.636175 valid's binary_logloss: 0.651137
[600] train's binary_logloss: 0.633356 valid's binary_logloss: 0.65085
[700] train's binary_logloss: 0.630829 valid's binary_logloss: 0.650912
Early stopping, best iteration is:
[631] train's binary_logloss: 0.63254 valid's binary_logloss: 0.650592
num_leaves, val_score: 0.650592: 35%|###5 | 7/20 [00:05<00:11, 1.17it/s][I 2020-09-27 04:44:12,795] Trial 13 finished with value: 0.6505921558008332 and parameters: {'num_leaves': 3}. Best is trial 13 with value: 0.6505921558008332.
num_leaves, val_score: 0.650592: 35%|###5 | 7/20 [00:05<00:11, 1.17it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004382 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.643838 valid's binary_logloss: 0.655968
[200] train's binary_logloss: 0.630606 valid's binary_logloss: 0.651656
[300] train's binary_logloss: 0.62092 valid's binary_logloss: 0.651649
Early stopping, best iteration is:
[275] train's binary_logloss: 0.623153 valid's binary_logloss: 0.651151
num_leaves, val_score: 0.650592: 40%|#### | 8/20 [00:06<00:09, 1.29it/s][I 2020-09-27 04:44:13,393] Trial 14 finished with value: 0.6511512029998484 and parameters: {'num_leaves': 6}. Best is trial 13 with value: 0.6505921558008332.
num_leaves, val_score: 0.650592: 40%|#### | 8/20 [00:06<00:09, 1.29it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005223 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.536652 valid's binary_logloss: 0.654767
Early stopping, best iteration is:
[88] train's binary_logloss: 0.548536 valid's binary_logloss: 0.654093
num_leaves, val_score: 0.650592: 45%|####5 | 9/20 [00:07<00:09, 1.12it/s][I 2020-09-27 04:44:14,546] Trial 15 finished with value: 0.6540930395271628 and parameters: {'num_leaves': 55}. Best is trial 13 with value: 0.6505921558008332.
num_leaves, val_score: 0.650592: 45%|####5 | 9/20 [00:07<00:09, 1.12it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000391 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.522391 valid's binary_logloss: 0.660187
Early stopping, best iteration is:
[56] train's binary_logloss: 0.573263 valid's binary_logloss: 0.659042
num_leaves, val_score: 0.650592: 50%|##### | 10/20 [00:08<00:07, 1.25it/s][I 2020-09-27 04:44:15,132] Trial 16 finished with value: 0.6590416618563378 and parameters: {'num_leaves': 63}. Best is trial 13 with value: 0.6505921558008332.
num_leaves, val_score: 0.650592: 50%|##### | 10/20 [00:08<00:07, 1.25it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000241 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.500962 valid's binary_logloss: 0.659121
Early stopping, best iteration is:
[60] train's binary_logloss: 0.552156 valid's binary_logloss: 0.658112
num_leaves, val_score: 0.650592: 55%|#####5 | 11/20 [00:08<00:06, 1.33it/s][I 2020-09-27 04:44:15,768] Trial 17 finished with value: 0.6581124834744749 and parameters: {'num_leaves': 77}. Best is trial 13 with value: 0.6505921558008332.
num_leaves, val_score: 0.650592: 55%|#####5 | 11/20 [00:08<00:06, 1.33it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000455 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.626446 valid's binary_logloss: 0.654342
[200] train's binary_logloss: 0.602897 valid's binary_logloss: 0.654305
Early stopping, best iteration is:
[128] train's binary_logloss: 0.619208 valid's binary_logloss: 0.653208
num_leaves, val_score: 0.650592: 60%|###### | 12/20 [00:09<00:05, 1.50it/s][I 2020-09-27 04:44:16,240] Trial 18 finished with value: 0.6532083579968055 and parameters: {'num_leaves': 12}. Best is trial 13 with value: 0.6505921558008332.
num_leaves, val_score: 0.650592: 60%|###### | 12/20 [00:09<00:05, 1.50it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000331 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.340384 valid's binary_logloss: 0.669675
Early stopping, best iteration is:
[43] train's binary_logloss: 0.483746 valid's binary_logloss: 0.660937
num_leaves, val_score: 0.650592: 65%|######5 | 13/20 [00:10<00:05, 1.31it/s][I 2020-09-27 04:44:17,237] Trial 19 finished with value: 0.6609367707387375 and parameters: {'num_leaves': 209}. Best is trial 13 with value: 0.6505921558008332.
num_leaves, val_score: 0.650592: 65%|######5 | 13/20 [00:10<00:05, 1.31it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000416 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.47136 valid's binary_logloss: 0.661857
Early stopping, best iteration is:
[66] train's binary_logloss: 0.520512 valid's binary_logloss: 0.659368
num_leaves, val_score: 0.650592: 70%|####### | 14/20 [00:11<00:05, 1.08it/s][I 2020-09-27 04:44:18,525] Trial 20 finished with value: 0.6593681790082808 and parameters: {'num_leaves': 95}. Best is trial 13 with value: 0.6505921558008332.
num_leaves, val_score: 0.650592: 70%|####### | 14/20 [00:11<00:05, 1.08it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000434 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.634417 valid's binary_logloss: 0.655116
[200] train's binary_logloss: 0.616494 valid's binary_logloss: 0.653099
[300] train's binary_logloss: 0.601536 valid's binary_logloss: 0.65265
Early stopping, best iteration is:
[250] train's binary_logloss: 0.608783 valid's binary_logloss: 0.651395
num_leaves, val_score: 0.650592: 75%|#######5 | 15/20 [00:12<00:04, 1.20it/s][I 2020-09-27 04:44:19,142] Trial 21 finished with value: 0.6513954160672126 and parameters: {'num_leaves': 9}. Best is trial 13 with value: 0.6505921558008332.
num_leaves, val_score: 0.650592: 75%|#######5 | 15/20 [00:12<00:04, 1.20it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000366 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.668969 valid's binary_logloss: 0.671303
[200] train's binary_logloss: 0.659328 valid's binary_logloss: 0.662192
[300] train's binary_logloss: 0.65408 valid's binary_logloss: 0.657049
[400] train's binary_logloss: 0.650963 valid's binary_logloss: 0.654645
[500] train's binary_logloss: 0.648956 valid's binary_logloss: 0.652696
[600] train's binary_logloss: 0.647608 valid's binary_logloss: 0.651783
[700] train's binary_logloss: 0.646665 valid's binary_logloss: 0.651143
[800] train's binary_logloss: 0.645958 valid's binary_logloss: 0.650652
[900] train's binary_logloss: 0.645405 valid's binary_logloss: 0.6504
[1000] train's binary_logloss: 0.644946 valid's binary_logloss: 0.650075
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.644946 valid's binary_logloss: 0.650075
num_leaves, val_score: 0.650075: 80%|######## | 16/20 [00:13<00:03, 1.05it/s][I 2020-09-27 04:44:20,393] Trial 22 finished with value: 0.6500754523975805 and parameters: {'num_leaves': 2}. Best is trial 22 with value: 0.6500754523975805.
num_leaves, val_score: 0.650075: 80%|######## | 16/20 [00:13<00:03, 1.05it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.005069 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.565299 valid's binary_logloss: 0.658702
Early stopping, best iteration is:
[60] train's binary_logloss: 0.597675 valid's binary_logloss: 0.657428
num_leaves, val_score: 0.650075: 85%|########5 | 17/20 [00:13<00:02, 1.23it/s][I 2020-09-27 04:44:20,871] Trial 23 finished with value: 0.6574278070760912 and parameters: {'num_leaves': 39}. Best is trial 22 with value: 0.6500754523975805.
num_leaves, val_score: 0.650075: 85%|########5 | 17/20 [00:13<00:02, 1.23it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000464 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.652156 valid's binary_logloss: 0.66018
[200] train's binary_logloss: 0.641398 valid's binary_logloss: 0.653863
[300] train's binary_logloss: 0.635014 valid's binary_logloss: 0.652565
[400] train's binary_logloss: 0.629702 valid's binary_logloss: 0.65154
[500] train's binary_logloss: 0.624786 valid's binary_logloss: 0.652002
Early stopping, best iteration is:
[409] train's binary_logloss: 0.629299 valid's binary_logloss: 0.651463
num_leaves, val_score: 0.650075: 90%|######### | 18/20 [00:15<00:01, 1.05it/s][I 2020-09-27 04:44:22,155] Trial 24 finished with value: 0.6514630998860335 and parameters: {'num_leaves': 4}. Best is trial 22 with value: 0.6500754523975805.
num_leaves, val_score: 0.650075: 90%|######### | 18/20 [00:15<00:01, 1.05it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000333 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.570374 valid's binary_logloss: 0.656731
Early stopping, best iteration is:
[72] train's binary_logloss: 0.590744 valid's binary_logloss: 0.655417
num_leaves, val_score: 0.650075: 95%|#########5| 19/20 [00:15<00:00, 1.23it/s][I 2020-09-27 04:44:22,639] Trial 25 finished with value: 0.655416641463825 and parameters: {'num_leaves': 37}. Best is trial 22 with value: 0.6500754523975805.
num_leaves, val_score: 0.650075: 95%|#########5| 19/20 [00:15<00:00, 1.23it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000693 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.455281 valid's binary_logloss: 0.66208
Early stopping, best iteration is:
[62] train's binary_logloss: 0.51632 valid's binary_logloss: 0.658037
num_leaves, val_score: 0.650075: 100%|##########| 20/20 [00:16<00:00, 1.26it/s][I 2020-09-27 04:44:23,396] Trial 26 finished with value: 0.6580367008399806 and parameters: {'num_leaves': 107}. Best is trial 22 with value: 0.6500754523975805.
num_leaves, val_score: 0.650075: 100%|##########| 20/20 [00:16<00:00, 1.21it/s]
bagging, val_score: 0.650075: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000433 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667722 valid's binary_logloss: 0.669888
[200] train's binary_logloss: 0.657469 valid's binary_logloss: 0.660998
[300] train's binary_logloss: 0.652189 valid's binary_logloss: 0.656002
[400] train's binary_logloss: 0.649008 valid's binary_logloss: 0.65329
[500] train's binary_logloss: 0.64724 valid's binary_logloss: 0.651649
[600] train's binary_logloss: 0.646196 valid's binary_logloss: 0.651202
[700] train's binary_logloss: 0.645282 valid's binary_logloss: 0.650767
[800] train's binary_logloss: 0.644572 valid's binary_logloss: 0.650399
[900] train's binary_logloss: 0.643859 valid's binary_logloss: 0.650238
[1000] train's binary_logloss: 0.643295 valid's binary_logloss: 0.649949
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.643295 valid's binary_logloss: 0.649949
bagging, val_score: 0.649949: 10%|# | 1/10 [00:01<00:12, 1.38s/it][I 2020-09-27 04:44:24,791] Trial 27 finished with value: 0.6499492453087289 and parameters: {'bagging_fraction': 0.6374753391797192, 'bagging_freq': 6}. Best is trial 27 with value: 0.6499492453087289.
bagging, val_score: 0.649949: 10%|# | 1/10 [00:01<00:12, 1.38s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000379 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667741 valid's binary_logloss: 0.670126
[200] train's binary_logloss: 0.657401 valid's binary_logloss: 0.660759
[300] train's binary_logloss: 0.652134 valid's binary_logloss: 0.655834
[400] train's binary_logloss: 0.649009 valid's binary_logloss: 0.653296
[500] train's binary_logloss: 0.647228 valid's binary_logloss: 0.651516
[600] train's binary_logloss: 0.646169 valid's binary_logloss: 0.65095
[700] train's binary_logloss: 0.645307 valid's binary_logloss: 0.650552
[800] train's binary_logloss: 0.644552 valid's binary_logloss: 0.650135
[900] train's binary_logloss: 0.643876 valid's binary_logloss: 0.650283
Early stopping, best iteration is:
[811] train's binary_logloss: 0.64445 valid's binary_logloss: 0.649941
bagging, val_score: 0.649941: 20%|## | 2/10 [00:03<00:11, 1.48s/it][I 2020-09-27 04:44:26,503] Trial 28 finished with value: 0.6499412034252128 and parameters: {'bagging_fraction': 0.6431039528688518, 'bagging_freq': 6}. Best is trial 28 with value: 0.6499412034252128.
bagging, val_score: 0.649941: 20%|## | 2/10 [00:03<00:11, 1.48s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000402 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844
[200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602
[300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818
[400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001
[500] train's binary_logloss: 0.647221 valid's binary_logloss: 0.651249
[600] train's binary_logloss: 0.646154 valid's binary_logloss: 0.650938
[700] train's binary_logloss: 0.645276 valid's binary_logloss: 0.650398
[800] train's binary_logloss: 0.644511 valid's binary_logloss: 0.65025
[900] train's binary_logloss: 0.643837 valid's binary_logloss: 0.650107
[1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721
bagging, val_score: 0.649721: 30%|### | 3/10 [00:04<00:10, 1.44s/it][I 2020-09-27 04:44:27,842] Trial 29 finished with value: 0.6497207980729593 and parameters: {'bagging_fraction': 0.6147135266937416, 'bagging_freq': 6}. Best is trial 29 with value: 0.6497207980729593.
bagging, val_score: 0.649721: 30%|### | 3/10 [00:04<00:10, 1.44s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000403 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.66762 valid's binary_logloss: 0.670222
[200] train's binary_logloss: 0.657378 valid's binary_logloss: 0.660905
[300] train's binary_logloss: 0.652057 valid's binary_logloss: 0.655995
[400] train's binary_logloss: 0.648925 valid's binary_logloss: 0.653365
[500] train's binary_logloss: 0.647153 valid's binary_logloss: 0.65154
[600] train's binary_logloss: 0.646118 valid's binary_logloss: 0.651221
[700] train's binary_logloss: 0.64522 valid's binary_logloss: 0.650848
[800] train's binary_logloss: 0.644499 valid's binary_logloss: 0.650374
[900] train's binary_logloss: 0.643818 valid's binary_logloss: 0.650522
Early stopping, best iteration is:
[811] train's binary_logloss: 0.644399 valid's binary_logloss: 0.650183
bagging, val_score: 0.649721: 40%|#### | 4/10 [00:05<00:08, 1.40s/it][I 2020-09-27 04:44:29,154] Trial 30 finished with value: 0.6501826485329506 and parameters: {'bagging_fraction': 0.6282514585296812, 'bagging_freq': 6}. Best is trial 29 with value: 0.6497207980729593.
bagging, val_score: 0.649721: 40%|#### | 4/10 [00:05<00:08, 1.40s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000425 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667636 valid's binary_logloss: 0.670402
[200] train's binary_logloss: 0.65741 valid's binary_logloss: 0.660721
[300] train's binary_logloss: 0.652135 valid's binary_logloss: 0.655888
[400] train's binary_logloss: 0.649005 valid's binary_logloss: 0.653394
[500] train's binary_logloss: 0.647265 valid's binary_logloss: 0.651817
[600] train's binary_logloss: 0.646169 valid's binary_logloss: 0.651429
[700] train's binary_logloss: 0.645289 valid's binary_logloss: 0.651273
Early stopping, best iteration is:
[625] train's binary_logloss: 0.645884 valid's binary_logloss: 0.650972
bagging, val_score: 0.649721: 50%|##### | 5/10 [00:07<00:06, 1.39s/it][I 2020-09-27 04:44:30,538] Trial 31 finished with value: 0.6509722145299361 and parameters: {'bagging_fraction': 0.625788067973422, 'bagging_freq': 6}. Best is trial 29 with value: 0.6497207980729593.
bagging, val_score: 0.649721: 50%|##### | 5/10 [00:07<00:06, 1.39s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000386 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.66771 valid's binary_logloss: 0.670672
[200] train's binary_logloss: 0.657453 valid's binary_logloss: 0.660978
[300] train's binary_logloss: 0.652108 valid's binary_logloss: 0.655884
[400] train's binary_logloss: 0.648979 valid's binary_logloss: 0.653214
[500] train's binary_logloss: 0.647206 valid's binary_logloss: 0.651492
[600] train's binary_logloss: 0.646131 valid's binary_logloss: 0.650889
[700] train's binary_logloss: 0.645274 valid's binary_logloss: 0.650764
[800] train's binary_logloss: 0.644537 valid's binary_logloss: 0.650318
[900] train's binary_logloss: 0.643867 valid's binary_logloss: 0.650313
Early stopping, best iteration is:
[811] train's binary_logloss: 0.644435 valid's binary_logloss: 0.650123
bagging, val_score: 0.649721: 60%|###### | 6/10 [00:08<00:05, 1.34s/it][I 2020-09-27 04:44:31,752] Trial 32 finished with value: 0.650122761267748 and parameters: {'bagging_fraction': 0.6308230415432386, 'bagging_freq': 6}. Best is trial 29 with value: 0.6497207980729593.
bagging, val_score: 0.649721: 60%|###### | 6/10 [00:08<00:05, 1.34s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000509 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.668475 valid's binary_logloss: 0.670977
[200] train's binary_logloss: 0.658545 valid's binary_logloss: 0.661641
[300] train's binary_logloss: 0.653145 valid's binary_logloss: 0.656461
[400] train's binary_logloss: 0.649945 valid's binary_logloss: 0.653659
[500] train's binary_logloss: 0.647957 valid's binary_logloss: 0.652033
[600] train's binary_logloss: 0.646663 valid's binary_logloss: 0.651209
[700] train's binary_logloss: 0.645743 valid's binary_logloss: 0.650593
[800] train's binary_logloss: 0.644998 valid's binary_logloss: 0.649891
[900] train's binary_logloss: 0.644366 valid's binary_logloss: 0.649745
[1000] train's binary_logloss: 0.643856 valid's binary_logloss: 0.649763
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.643856 valid's binary_logloss: 0.649763
bagging, val_score: 0.649721: 70%|####### | 7/10 [00:09<00:04, 1.41s/it][I 2020-09-27 04:44:33,330] Trial 33 finished with value: 0.649762914198675 and parameters: {'bagging_fraction': 0.881712034252823, 'bagging_freq': 6}. Best is trial 29 with value: 0.6497207980729593.
bagging, val_score: 0.649721: 70%|####### | 7/10 [00:09<00:04, 1.41s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001411 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.668676 valid's binary_logloss: 0.671455
[200] train's binary_logloss: 0.658729 valid's binary_logloss: 0.66187
[300] train's binary_logloss: 0.653368 valid's binary_logloss: 0.656513
[400] train's binary_logloss: 0.650201 valid's binary_logloss: 0.654227
[500] train's binary_logloss: 0.648173 valid's binary_logloss: 0.652343
[600] train's binary_logloss: 0.646868 valid's binary_logloss: 0.651436
[700] train's binary_logloss: 0.645933 valid's binary_logloss: 0.650838
[800] train's binary_logloss: 0.645188 valid's binary_logloss: 0.650137
[900] train's binary_logloss: 0.644607 valid's binary_logloss: 0.64995
[1000] train's binary_logloss: 0.644106 valid's binary_logloss: 0.649864
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.644106 valid's binary_logloss: 0.649864
bagging, val_score: 0.649721: 80%|######## | 8/10 [00:11<00:03, 1.51s/it][I 2020-09-27 04:44:35,075] Trial 34 finished with value: 0.6498641722707447 and parameters: {'bagging_fraction': 0.9222411833757059, 'bagging_freq': 6}. Best is trial 29 with value: 0.6497207980729593.
bagging, val_score: 0.649721: 80%|######## | 8/10 [00:11<00:03, 1.51s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000737 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.668902 valid's binary_logloss: 0.671648
[200] train's binary_logloss: 0.65921 valid's binary_logloss: 0.661972
[300] train's binary_logloss: 0.653923 valid's binary_logloss: 0.65701
[400] train's binary_logloss: 0.650779 valid's binary_logloss: 0.654505
[500] train's binary_logloss: 0.648744 valid's binary_logloss: 0.652567
[600] train's binary_logloss: 0.647377 valid's binary_logloss: 0.651548
[700] train's binary_logloss: 0.646422 valid's binary_logloss: 0.651081
[800] train's binary_logloss: 0.645708 valid's binary_logloss: 0.650642
[900] train's binary_logloss: 0.64514 valid's binary_logloss: 0.650392
[1000] train's binary_logloss: 0.644665 valid's binary_logloss: 0.650089
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.644665 valid's binary_logloss: 0.650089
bagging, val_score: 0.649721: 90%|######### | 9/10 [00:13<00:01, 1.50s/it][I 2020-09-27 04:44:36,546] Trial 35 finished with value: 0.6500890749178282 and parameters: {'bagging_fraction': 0.9922180140725776, 'bagging_freq': 2}. Best is trial 29 with value: 0.6497207980729593.
bagging, val_score: 0.649721: 90%|######### | 9/10 [00:13<00:01, 1.50s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000379 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.66868 valid's binary_logloss: 0.671198
[200] train's binary_logloss: 0.658732 valid's binary_logloss: 0.661411
[300] train's binary_logloss: 0.653393 valid's binary_logloss: 0.656638
[400] train's binary_logloss: 0.650183 valid's binary_logloss: 0.653991
[500] train's binary_logloss: 0.648188 valid's binary_logloss: 0.652694
[600] train's binary_logloss: 0.646874 valid's binary_logloss: 0.651422
[700] train's binary_logloss: 0.645949 valid's binary_logloss: 0.650978
[800] train's binary_logloss: 0.645235 valid's binary_logloss: 0.650342
[900] train's binary_logloss: 0.644629 valid's binary_logloss: 0.649845
[1000] train's binary_logloss: 0.644118 valid's binary_logloss: 0.649971
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.644118 valid's binary_logloss: 0.649971
bagging, val_score: 0.649721: 100%|##########| 10/10 [00:15<00:00, 1.61s/it][I 2020-09-27 04:44:38,424] Trial 36 finished with value: 0.6499714458134913 and parameters: {'bagging_fraction': 0.9261093121309201, 'bagging_freq': 7}. Best is trial 29 with value: 0.6497207980729593.
bagging, val_score: 0.649721: 100%|##########| 10/10 [00:15<00:00, 1.50s/it]
feature_fraction_stage2, val_score: 0.649721: 0%| | 0/3 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000513 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667513 valid's binary_logloss: 0.670285
[200] train's binary_logloss: 0.65728 valid's binary_logloss: 0.661217
[300] train's binary_logloss: 0.652006 valid's binary_logloss: 0.656051
[400] train's binary_logloss: 0.648934 valid's binary_logloss: 0.653472
[500] train's binary_logloss: 0.647192 valid's binary_logloss: 0.651779
[600] train's binary_logloss: 0.646099 valid's binary_logloss: 0.651503
[700] train's binary_logloss: 0.645255 valid's binary_logloss: 0.650916
[800] train's binary_logloss: 0.644508 valid's binary_logloss: 0.650651
[900] train's binary_logloss: 0.643868 valid's binary_logloss: 0.650537
Early stopping, best iteration is:
[811] train's binary_logloss: 0.644389 valid's binary_logloss: 0.650382
feature_fraction_stage2, val_score: 0.649721: 33%|###3 | 1/3 [00:01<00:02, 1.23s/it][I 2020-09-27 04:44:39,673] Trial 37 finished with value: 0.6503821331647123 and parameters: {'feature_fraction': 0.48000000000000004}. Best is trial 37 with value: 0.6503821331647123.
feature_fraction_stage2, val_score: 0.649721: 33%|###3 | 1/3 [00:01<00:02, 1.23s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000502 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667466 valid's binary_logloss: 0.670008
[200] train's binary_logloss: 0.657197 valid's binary_logloss: 0.660612
[300] train's binary_logloss: 0.651972 valid's binary_logloss: 0.655778
[400] train's binary_logloss: 0.648867 valid's binary_logloss: 0.653147
[500] train's binary_logloss: 0.647122 valid's binary_logloss: 0.651573
[600] train's binary_logloss: 0.646035 valid's binary_logloss: 0.650926
[700] train's binary_logloss: 0.645194 valid's binary_logloss: 0.650578
[800] train's binary_logloss: 0.644427 valid's binary_logloss: 0.650523
[900] train's binary_logloss: 0.643736 valid's binary_logloss: 0.650556
Early stopping, best iteration is:
[811] train's binary_logloss: 0.644312 valid's binary_logloss: 0.650291
feature_fraction_stage2, val_score: 0.649721: 67%|######6 | 2/3 [00:02<00:01, 1.29s/it][I 2020-09-27 04:44:41,097] Trial 38 finished with value: 0.6502911144681769 and parameters: {'feature_fraction': 0.41600000000000004}. Best is trial 38 with value: 0.6502911144681769.
feature_fraction_stage2, val_score: 0.649721: 67%|######6 | 2/3 [00:02<00:01, 1.29s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002477 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667513 valid's binary_logloss: 0.670285
[200] train's binary_logloss: 0.65728 valid's binary_logloss: 0.661217
[300] train's binary_logloss: 0.652006 valid's binary_logloss: 0.656051
[400] train's binary_logloss: 0.648934 valid's binary_logloss: 0.653472
[500] train's binary_logloss: 0.647192 valid's binary_logloss: 0.651779
[600] train's binary_logloss: 0.646099 valid's binary_logloss: 0.651503
[700] train's binary_logloss: 0.645255 valid's binary_logloss: 0.650916
[800] train's binary_logloss: 0.644508 valid's binary_logloss: 0.650651
[900] train's binary_logloss: 0.643868 valid's binary_logloss: 0.650537
Early stopping, best iteration is:
[811] train's binary_logloss: 0.644389 valid's binary_logloss: 0.650382
feature_fraction_stage2, val_score: 0.649721: 100%|##########| 3/3 [00:04<00:00, 1.33s/it][I 2020-09-27 04:44:42,520] Trial 39 finished with value: 0.6503821331647123 and parameters: {'feature_fraction': 0.44800000000000006}. Best is trial 38 with value: 0.6502911144681769.
feature_fraction_stage2, val_score: 0.649721: 100%|##########| 3/3 [00:04<00:00, 1.36s/it]
regularization_factors, val_score: 0.649721: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000379 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844
[200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602
[300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818
[400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001
[500] train's binary_logloss: 0.647221 valid's binary_logloss: 0.651249
[600] train's binary_logloss: 0.646154 valid's binary_logloss: 0.650938
[700] train's binary_logloss: 0.645276 valid's binary_logloss: 0.650398
[800] train's binary_logloss: 0.644511 valid's binary_logloss: 0.65025
[900] train's binary_logloss: 0.643837 valid's binary_logloss: 0.650107
[1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721
regularization_factors, val_score: 0.649721: 5%|5 | 1/20 [00:01<00:25, 1.35s/it][I 2020-09-27 04:44:43,887] Trial 40 finished with value: 0.6497208080532643 and parameters: {'lambda_l1': 4.606489682248236e-06, 'lambda_l2': 0.0004490091062700392}. Best is trial 40 with value: 0.6497208080532643.
regularization_factors, val_score: 0.649721: 5%|5 | 1/20 [00:01<00:25, 1.35s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000342 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844
[200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602
[300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818
[400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001
[500] train's binary_logloss: 0.647221 valid's binary_logloss: 0.651249
[600] train's binary_logloss: 0.646154 valid's binary_logloss: 0.650938
[700] train's binary_logloss: 0.645276 valid's binary_logloss: 0.650398
[800] train's binary_logloss: 0.644511 valid's binary_logloss: 0.65025
[900] train's binary_logloss: 0.643837 valid's binary_logloss: 0.650107
[1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721
regularization_factors, val_score: 0.649721: 10%|# | 2/20 [00:03<00:26, 1.50s/it][I 2020-09-27 04:44:45,731] Trial 41 finished with value: 0.6497208177282439 and parameters: {'lambda_l1': 3.367646783477069e-06, 'lambda_l2': 0.0006824013831978261}. Best is trial 40 with value: 0.6497208080532643.
regularization_factors, val_score: 0.649721: 10%|# | 2/20 [00:03<00:26, 1.50s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000463 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844
[200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602
[300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818
[400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001
[500] train's binary_logloss: 0.647221 valid's binary_logloss: 0.651249
[600] train's binary_logloss: 0.646154 valid's binary_logloss: 0.650938
[700] train's binary_logloss: 0.645276 valid's binary_logloss: 0.650398
[800] train's binary_logloss: 0.644511 valid's binary_logloss: 0.65025
[900] train's binary_logloss: 0.643837 valid's binary_logloss: 0.650107
[1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721
regularization_factors, val_score: 0.649721: 15%|#5 | 3/20 [00:04<00:24, 1.44s/it][I 2020-09-27 04:44:47,018] Trial 42 finished with value: 0.6497208110181181 and parameters: {'lambda_l1': 1.2594003820953281e-06, 'lambda_l2': 0.0005901706717435754}. Best is trial 40 with value: 0.6497208080532643.
regularization_factors, val_score: 0.649721: 15%|#5 | 3/20 [00:04<00:24, 1.44s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000389 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844
[200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602
[300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818
[400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001
[500] train's binary_logloss: 0.647221 valid's binary_logloss: 0.651249
[600] train's binary_logloss: 0.646154 valid's binary_logloss: 0.650938
[700] train's binary_logloss: 0.645276 valid's binary_logloss: 0.650398
[800] train's binary_logloss: 0.644511 valid's binary_logloss: 0.65025
[900] train's binary_logloss: 0.643838 valid's binary_logloss: 0.650107
[1000] train's binary_logloss: 0.643266 valid's binary_logloss: 0.649721
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.643266 valid's binary_logloss: 0.649721
regularization_factors, val_score: 0.649721: 20%|## | 4/20 [00:05<00:22, 1.38s/it][I 2020-09-27 04:44:48,282] Trial 43 finished with value: 0.6497208467241731 and parameters: {'lambda_l1': 1.7272310002902818e-06, 'lambda_l2': 0.0005516895364430477}. Best is trial 40 with value: 0.6497208080532643.
regularization_factors, val_score: 0.649721: 20%|## | 4/20 [00:05<00:22, 1.38s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000424 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844
[200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602
[300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818
[400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001
[500] train's binary_logloss: 0.647221 valid's binary_logloss: 0.651249
[600] train's binary_logloss: 0.646154 valid's binary_logloss: 0.650938
[700] train's binary_logloss: 0.645276 valid's binary_logloss: 0.650398
[800] train's binary_logloss: 0.644512 valid's binary_logloss: 0.65025
[900] train's binary_logloss: 0.643838 valid's binary_logloss: 0.650107
[1000] train's binary_logloss: 0.643266 valid's binary_logloss: 0.649721
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.643266 valid's binary_logloss: 0.649721
regularization_factors, val_score: 0.649721: 25%|##5 | 5/20 [00:07<00:22, 1.49s/it][I 2020-09-27 04:44:50,019] Trial 44 finished with value: 0.6497209472919824 and parameters: {'lambda_l1': 1.66122924086294e-06, 'lambda_l2': 0.0004859937346429329}. Best is trial 40 with value: 0.6497208080532643.
regularization_factors, val_score: 0.649721: 25%|##5 | 5/20 [00:07<00:22, 1.49s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000580 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844
[200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602
[300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818
[400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001
[500] train's binary_logloss: 0.647221 valid's binary_logloss: 0.651249
[600] train's binary_logloss: 0.646154 valid's binary_logloss: 0.650938
[700] train's binary_logloss: 0.645276 valid's binary_logloss: 0.650398
[800] train's binary_logloss: 0.64451 valid's binary_logloss: 0.65025
[900] train's binary_logloss: 0.643837 valid's binary_logloss: 0.650107
[1000] train's binary_logloss: 0.643264 valid's binary_logloss: 0.649721
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.643264 valid's binary_logloss: 0.649721
regularization_factors, val_score: 0.649721: 30%|### | 6/20 [00:08<00:20, 1.45s/it][I 2020-09-27 04:44:51,378] Trial 45 finished with value: 0.6497207058626523 and parameters: {'lambda_l1': 3.2445067392487893e-06, 'lambda_l2': 0.0004261866691401461}. Best is trial 45 with value: 0.6497207058626523.
regularization_factors, val_score: 0.649721: 30%|### | 6/20 [00:08<00:20, 1.45s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000393 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844
[200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602
[300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818
[400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001
[500] train's binary_logloss: 0.647221 valid's binary_logloss: 0.651249
[600] train's binary_logloss: 0.646154 valid's binary_logloss: 0.650938
[700] train's binary_logloss: 0.645276 valid's binary_logloss: 0.650398
[800] train's binary_logloss: 0.644511 valid's binary_logloss: 0.65025
[900] train's binary_logloss: 0.643837 valid's binary_logloss: 0.650107
[1000] train's binary_logloss: 0.643264 valid's binary_logloss: 0.649721
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.643264 valid's binary_logloss: 0.649721
regularization_factors, val_score: 0.649721: 35%|###5 | 7/20 [00:10<00:19, 1.53s/it][I 2020-09-27 04:44:53,078] Trial 46 finished with value: 0.6497207182286581 and parameters: {'lambda_l1': 3.850005827927253e-06, 'lambda_l2': 0.0005152754958015667}. Best is trial 45 with value: 0.6497207058626523.
regularization_factors, val_score: 0.649721: 35%|###5 | 7/20 [00:10<00:19, 1.53s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000406 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844
[200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602
[300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818
[400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001
[500] train's binary_logloss: 0.647221 valid's binary_logloss: 0.651249
[600] train's binary_logloss: 0.646154 valid's binary_logloss: 0.650938
[700] train's binary_logloss: 0.645276 valid's binary_logloss: 0.650398
[800] train's binary_logloss: 0.644511 valid's binary_logloss: 0.65025
[900] train's binary_logloss: 0.643837 valid's binary_logloss: 0.650107
[1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721
regularization_factors, val_score: 0.649721: 40%|#### | 8/20 [00:11<00:17, 1.49s/it][I 2020-09-27 04:44:54,488] Trial 47 finished with value: 0.6497208048936114 and parameters: {'lambda_l1': 6.216594964021316e-06, 'lambda_l2': 0.0003007341715934148}. Best is trial 45 with value: 0.6497207058626523.
regularization_factors, val_score: 0.649721: 40%|#### | 8/20 [00:11<00:17, 1.49s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000538 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844
[200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602
[300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818
[400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001
[500] train's binary_logloss: 0.647226 valid's binary_logloss: 0.65118
[600] train's binary_logloss: 0.64616 valid's binary_logloss: 0.650744
[700] train's binary_logloss: 0.645279 valid's binary_logloss: 0.650352
[800] train's binary_logloss: 0.644543 valid's binary_logloss: 0.650069
[900] train's binary_logloss: 0.643859 valid's binary_logloss: 0.649948
[1000] train's binary_logloss: 0.643284 valid's binary_logloss: 0.649627
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.643284 valid's binary_logloss: 0.649627
regularization_factors, val_score: 0.649627: 45%|####5 | 9/20 [00:13<00:15, 1.45s/it][I 2020-09-27 04:44:55,828] Trial 48 finished with value: 0.6496273811877121 and parameters: {'lambda_l1': 0.0008017384871250849, 'lambda_l2': 2.2446245169734426e-07}. Best is trial 48 with value: 0.6496273811877121.
regularization_factors, val_score: 0.649627: 45%|####5 | 9/20 [00:13<00:15, 1.45s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000443 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667565 valid's binary_logloss: 0.669845
[200] train's binary_logloss: 0.657208 valid's binary_logloss: 0.660603
[300] train's binary_logloss: 0.651961 valid's binary_logloss: 0.65582
[400] train's binary_logloss: 0.648962 valid's binary_logloss: 0.65306
[500] train's binary_logloss: 0.647204 valid's binary_logloss: 0.651318
[600] train's binary_logloss: 0.646126 valid's binary_logloss: 0.650793
[700] train's binary_logloss: 0.645262 valid's binary_logloss: 0.650639
[800] train's binary_logloss: 0.644495 valid's binary_logloss: 0.650417
[900] train's binary_logloss: 0.643822 valid's binary_logloss: 0.650127
[1000] train's binary_logloss: 0.643242 valid's binary_logloss: 0.649774
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.643242 valid's binary_logloss: 0.649774
regularization_factors, val_score: 0.649627: 50%|##### | 10/20 [00:15<00:15, 1.53s/it][I 2020-09-27 04:44:57,545] Trial 49 finished with value: 0.6497744578765494 and parameters: {'lambda_l1': 0.012286863044744575, 'lambda_l2': 1.2600943736007061e-08}. Best is trial 48 with value: 0.6496273811877121.
regularization_factors, val_score: 0.649627: 50%|##### | 10/20 [00:15<00:15, 1.53s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000454 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667583 valid's binary_logloss: 0.6698
[200] train's binary_logloss: 0.657274 valid's binary_logloss: 0.66074
[300] train's binary_logloss: 0.652046 valid's binary_logloss: 0.655766
[400] train's binary_logloss: 0.649051 valid's binary_logloss: 0.65315
[500] train's binary_logloss: 0.647312 valid's binary_logloss: 0.651523
[600] train's binary_logloss: 0.646232 valid's binary_logloss: 0.651124
[700] train's binary_logloss: 0.645397 valid's binary_logloss: 0.650768
[800] train's binary_logloss: 0.644668 valid's binary_logloss: 0.650466
[900] train's binary_logloss: 0.643987 valid's binary_logloss: 0.650261
[1000] train's binary_logloss: 0.643431 valid's binary_logloss: 0.649835
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.643431 valid's binary_logloss: 0.649835
regularization_factors, val_score: 0.649627: 55%|#####5 | 11/20 [00:16<00:13, 1.47s/it][I 2020-09-27 04:44:58,890] Trial 50 finished with value: 0.6498345397919513 and parameters: {'lambda_l1': 0.0005572386578667407, 'lambda_l2': 2.4208863399594898}. Best is trial 48 with value: 0.6496273811877121.
regularization_factors, val_score: 0.649627: 55%|#####5 | 11/20 [00:16<00:13, 1.47s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000464 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844
[200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602
[300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818
[400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001
[500] train's binary_logloss: 0.647221 valid's binary_logloss: 0.651249
[600] train's binary_logloss: 0.646154 valid's binary_logloss: 0.650938
[700] train's binary_logloss: 0.645276 valid's binary_logloss: 0.650398
[800] train's binary_logloss: 0.644511 valid's binary_logloss: 0.65025
[900] train's binary_logloss: 0.643837 valid's binary_logloss: 0.650107
[1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721
regularization_factors, val_score: 0.649627: 60%|###### | 12/20 [00:18<00:12, 1.52s/it][I 2020-09-27 04:45:00,537] Trial 51 finished with value: 0.6497208015259045 and parameters: {'lambda_l1': 8.699166692204462e-05, 'lambda_l2': 3.087574877344173e-06}. Best is trial 48 with value: 0.6496273811877121.
regularization_factors, val_score: 0.649627: 60%|###### | 12/20 [00:18<00:12, 1.52s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002115 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844
[200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602
[300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818
[400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001
[500] train's binary_logloss: 0.647221 valid's binary_logloss: 0.651249
[600] train's binary_logloss: 0.646154 valid's binary_logloss: 0.650938
[700] train's binary_logloss: 0.645276 valid's binary_logloss: 0.650398
[800] train's binary_logloss: 0.644511 valid's binary_logloss: 0.65025
[900] train's binary_logloss: 0.643837 valid's binary_logloss: 0.650107
[1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721
regularization_factors, val_score: 0.649627: 65%|######5 | 13/20 [00:19<00:10, 1.50s/it][I 2020-09-27 04:45:01,964] Trial 52 finished with value: 0.6497208051737818 and parameters: {'lambda_l1': 0.00018230426317109586, 'lambda_l2': 4.548397960595403e-07}. Best is trial 48 with value: 0.6496273811877121.
regularization_factors, val_score: 0.649627: 65%|######5 | 13/20 [00:19<00:10, 1.50s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000506 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844
[200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602
[300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818
[400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001
[500] train's binary_logloss: 0.647221 valid's binary_logloss: 0.651249
[600] train's binary_logloss: 0.646154 valid's binary_logloss: 0.650938
[700] train's binary_logloss: 0.645276 valid's binary_logloss: 0.650398
[800] train's binary_logloss: 0.644511 valid's binary_logloss: 0.65025
[900] train's binary_logloss: 0.643837 valid's binary_logloss: 0.650107
[1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721
regularization_factors, val_score: 0.649627: 70%|####### | 14/20 [00:20<00:08, 1.43s/it][I 2020-09-27 04:45:03,256] Trial 53 finished with value: 0.6497208043193139 and parameters: {'lambda_l1': 0.00015987949714084808, 'lambda_l2': 1.645060058121504e-06}. Best is trial 48 with value: 0.6496273811877121.
regularization_factors, val_score: 0.649627: 70%|####### | 14/20 [00:20<00:08, 1.43s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000478 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844
[200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602
[300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818
[400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001
[500] train's binary_logloss: 0.647221 valid's binary_logloss: 0.651249
[600] train's binary_logloss: 0.646154 valid's binary_logloss: 0.650938
[700] train's binary_logloss: 0.645276 valid's binary_logloss: 0.650398
[800] train's binary_logloss: 0.644511 valid's binary_logloss: 0.650251
[900] train's binary_logloss: 0.643837 valid's binary_logloss: 0.650107
[1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721
regularization_factors, val_score: 0.649627: 75%|#######5 | 15/20 [00:22<00:07, 1.54s/it][I 2020-09-27 04:45:05,051] Trial 54 finished with value: 0.6497208216187388 and parameters: {'lambda_l1': 0.0006042412420761168, 'lambda_l2': 1.2746381508914564e-06}. Best is trial 48 with value: 0.6496273811877121.
regularization_factors, val_score: 0.649627: 75%|#######5 | 15/20 [00:22<00:07, 1.54s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000393 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844
[200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602
[300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818
[400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001
[500] train's binary_logloss: 0.647221 valid's binary_logloss: 0.651249
[600] train's binary_logloss: 0.646154 valid's binary_logloss: 0.650938
[700] train's binary_logloss: 0.645276 valid's binary_logloss: 0.650398
[800] train's binary_logloss: 0.644511 valid's binary_logloss: 0.65025
[900] train's binary_logloss: 0.643837 valid's binary_logloss: 0.650107
[1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721
regularization_factors, val_score: 0.649627: 80%|######## | 16/20 [00:23<00:05, 1.50s/it][I 2020-09-27 04:45:06,437] Trial 55 finished with value: 0.6497208030993801 and parameters: {'lambda_l1': 0.00012765660632242594, 'lambda_l2': 3.053198787844645e-06}. Best is trial 48 with value: 0.6496273811877121.
regularization_factors, val_score: 0.649627: 80%|######## | 16/20 [00:23<00:05, 1.50s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.005265 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844
[200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602
[300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818
[400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001
[500] train's binary_logloss: 0.647221 valid's binary_logloss: 0.651249
[600] train's binary_logloss: 0.646154 valid's binary_logloss: 0.650938
[700] train's binary_logloss: 0.645276 valid's binary_logloss: 0.650398
[800] train's binary_logloss: 0.644511 valid's binary_logloss: 0.65025
[900] train's binary_logloss: 0.643837 valid's binary_logloss: 0.650107
[1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721
regularization_factors, val_score: 0.649627: 85%|########5 | 17/20 [00:25<00:04, 1.47s/it][I 2020-09-27 04:45:07,858] Trial 56 finished with value: 0.6497208043349164 and parameters: {'lambda_l1': 4.556413554561587e-05, 'lambda_l2': 1.5942089823943423e-05}. Best is trial 48 with value: 0.6496273811877121.
regularization_factors, val_score: 0.649627: 85%|########5 | 17/20 [00:25<00:04, 1.47s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000436 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667565 valid's binary_logloss: 0.669845
[200] train's binary_logloss: 0.657208 valid's binary_logloss: 0.660602
[300] train's binary_logloss: 0.651961 valid's binary_logloss: 0.655819
[400] train's binary_logloss: 0.648963 valid's binary_logloss: 0.65306
[500] train's binary_logloss: 0.647204 valid's binary_logloss: 0.651318
[600] train's binary_logloss: 0.646126 valid's binary_logloss: 0.650793
[700] train's binary_logloss: 0.645263 valid's binary_logloss: 0.650638
[800] train's binary_logloss: 0.644496 valid's binary_logloss: 0.650417
[900] train's binary_logloss: 0.643824 valid's binary_logloss: 0.650127
[1000] train's binary_logloss: 0.643243 valid's binary_logloss: 0.649775
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.643243 valid's binary_logloss: 0.649775
regularization_factors, val_score: 0.649627: 90%|######### | 18/20 [00:26<00:03, 1.52s/it][I 2020-09-27 04:45:09,475] Trial 57 finished with value: 0.6497749853707365 and parameters: {'lambda_l1': 9.790223265016275e-08, 'lambda_l2': 0.04135439527004478}. Best is trial 48 with value: 0.6496273811877121.
regularization_factors, val_score: 0.649627: 90%|######### | 18/20 [00:26<00:03, 1.52s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000415 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669845
[200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602
[300] train's binary_logloss: 0.65196 valid's binary_logloss: 0.655819
[400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001
[500] train's binary_logloss: 0.647227 valid's binary_logloss: 0.65118
[600] train's binary_logloss: 0.64616 valid's binary_logloss: 0.650744
[700] train's binary_logloss: 0.64528 valid's binary_logloss: 0.650352
[800] train's binary_logloss: 0.644543 valid's binary_logloss: 0.65007
[900] train's binary_logloss: 0.64386 valid's binary_logloss: 0.649948
[1000] train's binary_logloss: 0.643285 valid's binary_logloss: 0.649628
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.643285 valid's binary_logloss: 0.649628
regularization_factors, val_score: 0.649627: 95%|#########5| 19/20 [00:28<00:01, 1.46s/it][I 2020-09-27 04:45:10,790] Trial 58 finished with value: 0.6496276632112322 and parameters: {'lambda_l1': 0.0038892159453483817, 'lambda_l2': 3.162000398193428e-08}. Best is trial 48 with value: 0.6496273811877121.
regularization_factors, val_score: 0.649627: 95%|#########5| 19/20 [00:28<00:01, 1.46s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000477 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667565 valid's binary_logloss: 0.669846
[200] train's binary_logloss: 0.657208 valid's binary_logloss: 0.660603
[300] train's binary_logloss: 0.651961 valid's binary_logloss: 0.65582
[400] train's binary_logloss: 0.648963 valid's binary_logloss: 0.65306
[500] train's binary_logloss: 0.647204 valid's binary_logloss: 0.651318
[600] train's binary_logloss: 0.646126 valid's binary_logloss: 0.650793
[700] train's binary_logloss: 0.645262 valid's binary_logloss: 0.650639
[800] train's binary_logloss: 0.644495 valid's binary_logloss: 0.650418
[900] train's binary_logloss: 0.643823 valid's binary_logloss: 0.650127
[1000] train's binary_logloss: 0.643243 valid's binary_logloss: 0.649775
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.643243 valid's binary_logloss: 0.649775
regularization_factors, val_score: 0.649627: 100%|##########| 20/20 [00:30<00:00, 1.59s/it][I 2020-09-27 04:45:12,683] Trial 59 finished with value: 0.6497745906504417 and parameters: {'lambda_l1': 0.012927722016684172, 'lambda_l2': 1.3099842616085507e-08}. Best is trial 48 with value: 0.6496273811877121.
regularization_factors, val_score: 0.649627: 100%|##########| 20/20 [00:30<00:00, 1.51s/it]
min_data_in_leaf, val_score: 0.649627: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000494 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844
[200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602
[300] train's binary_logloss: 0.652009 valid's binary_logloss: 0.655779
[400] train's binary_logloss: 0.649064 valid's binary_logloss: 0.653022
[500] train's binary_logloss: 0.647474 valid's binary_logloss: 0.651157
[600] train's binary_logloss: 0.646465 valid's binary_logloss: 0.650811
[700] train's binary_logloss: 0.645712 valid's binary_logloss: 0.650417
[800] train's binary_logloss: 0.645041 valid's binary_logloss: 0.650168
[900] train's binary_logloss: 0.644428 valid's binary_logloss: 0.650337
Early stopping, best iteration is:
[811] train's binary_logloss: 0.644942 valid's binary_logloss: 0.65
min_data_in_leaf, val_score: 0.649627: 20%|## | 1/5 [00:01<00:05, 1.26s/it][I 2020-09-27 04:45:13,958] Trial 60 finished with value: 0.6500004379994676 and parameters: {'min_child_samples': 100}. Best is trial 60 with value: 0.6500004379994676.
min_data_in_leaf, val_score: 0.649627: 20%|## | 1/5 [00:01<00:05, 1.26s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000400 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844
[200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602
[300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818
[400] train's binary_logloss: 0.648932 valid's binary_logloss: 0.652828
[500] train's binary_logloss: 0.647269 valid's binary_logloss: 0.651176
[600] train's binary_logloss: 0.646216 valid's binary_logloss: 0.650912
[700] train's binary_logloss: 0.64537 valid's binary_logloss: 0.650588
[800] train's binary_logloss: 0.644654 valid's binary_logloss: 0.65043
[900] train's binary_logloss: 0.643973 valid's binary_logloss: 0.650308
Early stopping, best iteration is:
[811] train's binary_logloss: 0.644548 valid's binary_logloss: 0.650197
min_data_in_leaf, val_score: 0.649627: 40%|#### | 2/5 [00:02<00:03, 1.26s/it][I 2020-09-27 04:45:15,217] Trial 61 finished with value: 0.6501970617141319 and parameters: {'min_child_samples': 50}. Best is trial 60 with value: 0.6500004379994676.
min_data_in_leaf, val_score: 0.649627: 40%|#### | 2/5 [00:02<00:03, 1.26s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000404 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844
[200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602
[300] train's binary_logloss: 0.65195 valid's binary_logloss: 0.65587
[400] train's binary_logloss: 0.648925 valid's binary_logloss: 0.65304
[500] train's binary_logloss: 0.647138 valid's binary_logloss: 0.651221
[600] train's binary_logloss: 0.646029 valid's binary_logloss: 0.650864
[700] train's binary_logloss: 0.645079 valid's binary_logloss: 0.650534
[800] train's binary_logloss: 0.644313 valid's binary_logloss: 0.650116
[900] train's binary_logloss: 0.643631 valid's binary_logloss: 0.650098
Early stopping, best iteration is:
[811] train's binary_logloss: 0.644204 valid's binary_logloss: 0.649909
min_data_in_leaf, val_score: 0.649627: 60%|###### | 3/5 [00:04<00:02, 1.36s/it][I 2020-09-27 04:45:16,819] Trial 62 finished with value: 0.6499094175863028 and parameters: {'min_child_samples': 5}. Best is trial 62 with value: 0.6499094175863028.
min_data_in_leaf, val_score: 0.649627: 60%|###### | 3/5 [00:04<00:02, 1.36s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000374 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844
[200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602
[300] train's binary_logloss: 0.65195 valid's binary_logloss: 0.65587
[400] train's binary_logloss: 0.648927 valid's binary_logloss: 0.653037
[500] train's binary_logloss: 0.647134 valid's binary_logloss: 0.651144
[600] train's binary_logloss: 0.646045 valid's binary_logloss: 0.650746
[700] train's binary_logloss: 0.64509 valid's binary_logloss: 0.650337
[800] train's binary_logloss: 0.644337 valid's binary_logloss: 0.650057
[900] train's binary_logloss: 0.643666 valid's binary_logloss: 0.649979
Early stopping, best iteration is:
[811] train's binary_logloss: 0.64422 valid's binary_logloss: 0.649764
min_data_in_leaf, val_score: 0.649627: 80%|######## | 4/5 [00:05<00:01, 1.32s/it][I 2020-09-27 04:45:18,047] Trial 63 finished with value: 0.649763563681734 and parameters: {'min_child_samples': 10}. Best is trial 63 with value: 0.649763563681734.
min_data_in_leaf, val_score: 0.649627: 80%|######## | 4/5 [00:05<00:01, 1.32s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000400 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844
[200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602
[300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818
[400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001
[500] train's binary_logloss: 0.647226 valid's binary_logloss: 0.65118
[600] train's binary_logloss: 0.646171 valid's binary_logloss: 0.650777
[700] train's binary_logloss: 0.645308 valid's binary_logloss: 0.650525
[800] train's binary_logloss: 0.644563 valid's binary_logloss: 0.650372
[900] train's binary_logloss: 0.643935 valid's binary_logloss: 0.650147
[1000] train's binary_logloss: 0.643333 valid's binary_logloss: 0.649925
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.643333 valid's binary_logloss: 0.649925
min_data_in_leaf, val_score: 0.649627: 100%|##########| 5/5 [00:06<00:00, 1.32s/it][I 2020-09-27 04:45:19,357] Trial 64 finished with value: 0.6499246052491834 and parameters: {'min_child_samples': 25}. Best is trial 63 with value: 0.649763563681734.
min_data_in_leaf, val_score: 0.649627: 100%|##########| 5/5 [00:06<00:00, 1.33s/it]
Fold : 7
[I 2020-09-27 04:45:19,501] A new study created in memory with name: no-name-a7907594-6487-49b5-9a75-d6ef7c2f3b45
feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008857 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.576754 valid's binary_logloss: 0.65897
Early stopping, best iteration is:
[69] train's binary_logloss: 0.59731 valid's binary_logloss: 0.657355
feature_fraction, val_score: 0.657355: 14%|#4 | 1/7 [00:00<00:05, 1.05it/s][I 2020-09-27 04:45:20,471] Trial 0 finished with value: 0.6573548252031896 and parameters: {'feature_fraction': 0.6}. Best is trial 0 with value: 0.6573548252031896.
feature_fraction, val_score: 0.657355: 14%|#4 | 1/7 [00:00<00:05, 1.05it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000492 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.578741 valid's binary_logloss: 0.657569
[200] train's binary_logloss: 0.527428 valid's binary_logloss: 0.66217
Early stopping, best iteration is:
[101] train's binary_logloss: 0.578076 valid's binary_logloss: 0.657517
feature_fraction, val_score: 0.657355: 29%|##8 | 2/7 [00:01<00:04, 1.19it/s][I 2020-09-27 04:45:21,053] Trial 1 finished with value: 0.6575168666853799 and parameters: {'feature_fraction': 0.5}. Best is trial 0 with value: 0.6573548252031896.
feature_fraction, val_score: 0.657355: 29%|##8 | 2/7 [00:01<00:04, 1.19it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000966 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.572623 valid's binary_logloss: 0.65727
Early stopping, best iteration is:
[81] train's binary_logloss: 0.585015 valid's binary_logloss: 0.656834
feature_fraction, val_score: 0.656834: 43%|####2 | 3/7 [00:02<00:03, 1.31it/s][I 2020-09-27 04:45:21,630] Trial 2 finished with value: 0.6568343510931819 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 2 with value: 0.6568343510931819.
feature_fraction, val_score: 0.656834: 43%|####2 | 3/7 [00:02<00:03, 1.31it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008151 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.573507 valid's binary_logloss: 0.657017
[200] train's binary_logloss: 0.517816 valid's binary_logloss: 0.65931
Early stopping, best iteration is:
[131] train's binary_logloss: 0.55462 valid's binary_logloss: 0.655904
feature_fraction, val_score: 0.655904: 57%|#####7 | 4/7 [00:02<00:02, 1.36it/s][I 2020-09-27 04:45:22,306] Trial 3 finished with value: 0.6559041635016774 and parameters: {'feature_fraction': 0.8}. Best is trial 3 with value: 0.6559041635016774.
feature_fraction, val_score: 0.655904: 57%|#####7 | 4/7 [00:02<00:02, 1.36it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005133 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.576598 valid's binary_logloss: 0.657946
Early stopping, best iteration is:
[91] train's binary_logloss: 0.582318 valid's binary_logloss: 0.656196
feature_fraction, val_score: 0.655904: 71%|#######1 | 5/7 [00:03<00:01, 1.49it/s][I 2020-09-27 04:45:22,831] Trial 4 finished with value: 0.6561957860921139 and parameters: {'feature_fraction': 0.7}. Best is trial 3 with value: 0.6559041635016774.
feature_fraction, val_score: 0.655904: 71%|#######1 | 5/7 [00:03<00:01, 1.49it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000955 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.570187 valid's binary_logloss: 0.658113
Early stopping, best iteration is:
[86] train's binary_logloss: 0.579352 valid's binary_logloss: 0.657006
feature_fraction, val_score: 0.655904: 86%|########5 | 6/7 [00:04<00:00, 1.27it/s][I 2020-09-27 04:45:23,895] Trial 5 finished with value: 0.6570061744677883 and parameters: {'feature_fraction': 1.0}. Best is trial 3 with value: 0.6559041635016774.
feature_fraction, val_score: 0.655904: 86%|########5 | 6/7 [00:04<00:00, 1.27it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000543 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.582693 valid's binary_logloss: 0.657081
[200] train's binary_logloss: 0.532209 valid's binary_logloss: 0.662404
Early stopping, best iteration is:
[100] train's binary_logloss: 0.582693 valid's binary_logloss: 0.657081
feature_fraction, val_score: 0.655904: 100%|##########| 7/7 [00:04<00:00, 1.39it/s][I 2020-09-27 04:45:24,444] Trial 6 finished with value: 0.6570805680910938 and parameters: {'feature_fraction': 0.4}. Best is trial 3 with value: 0.6559041635016774.
feature_fraction, val_score: 0.655904: 100%|##########| 7/7 [00:04<00:00, 1.42it/s]
num_leaves, val_score: 0.655904: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004860 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.466323 valid's binary_logloss: 0.658116
Early stopping, best iteration is:
[54] train's binary_logloss: 0.534041 valid's binary_logloss: 0.655259
num_leaves, val_score: 0.655259: 5%|5 | 1/20 [00:00<00:14, 1.35it/s][I 2020-09-27 04:45:25,204] Trial 7 finished with value: 0.6552590163821485 and parameters: {'num_leaves': 85}. Best is trial 7 with value: 0.6552590163821485.
num_leaves, val_score: 0.655259: 5%|5 | 1/20 [00:00<00:14, 1.35it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000895 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.540475 valid's binary_logloss: 0.662426
Early stopping, best iteration is:
[83] train's binary_logloss: 0.556094 valid's binary_logloss: 0.65974
num_leaves, val_score: 0.655259: 10%|# | 2/20 [00:01<00:12, 1.40it/s][I 2020-09-27 04:45:25,845] Trial 8 finished with value: 0.6597395208090874 and parameters: {'num_leaves': 46}. Best is trial 7 with value: 0.6552590163821485.
num_leaves, val_score: 0.655259: 10%|# | 2/20 [00:01<00:12, 1.40it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004966 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.60271 valid's binary_logloss: 0.656639
[200] train's binary_logloss: 0.564039 valid's binary_logloss: 0.658129
Early stopping, best iteration is:
[108] train's binary_logloss: 0.59936 valid's binary_logloss: 0.65622
num_leaves, val_score: 0.655259: 15%|#5 | 3/20 [00:01<00:11, 1.53it/s][I 2020-09-27 04:45:26,358] Trial 9 finished with value: 0.6562201154063112 and parameters: {'num_leaves': 19}. Best is trial 7 with value: 0.6552590163821485.
num_leaves, val_score: 0.655259: 15%|#5 | 3/20 [00:01<00:11, 1.53it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000795 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.314803 valid's binary_logloss: 0.675418
Early stopping, best iteration is:
[17] train's binary_logloss: 0.568221 valid's binary_logloss: 0.663895
num_leaves, val_score: 0.655259: 20%|## | 4/20 [00:03<00:15, 1.05it/s][I 2020-09-27 04:45:28,019] Trial 10 finished with value: 0.6638954298043613 and parameters: {'num_leaves': 201}. Best is trial 7 with value: 0.6552590163821485.
num_leaves, val_score: 0.655259: 20%|## | 4/20 [00:03<00:15, 1.05it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001158 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.415347 valid's binary_logloss: 0.669083
Early stopping, best iteration is:
[25] train's binary_logloss: 0.574128 valid's binary_logloss: 0.660544
num_leaves, val_score: 0.655259: 25%|##5 | 5/20 [00:04<00:13, 1.10it/s][I 2020-09-27 04:45:28,824] Trial 11 finished with value: 0.6605435697983096 and parameters: {'num_leaves': 118}. Best is trial 7 with value: 0.6552590163821485.
num_leaves, val_score: 0.655259: 25%|##5 | 5/20 [00:04<00:13, 1.10it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000788 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.424268 valid's binary_logloss: 0.667432
Early stopping, best iteration is:
[24] train's binary_logloss: 0.581498 valid's binary_logloss: 0.659094
num_leaves, val_score: 0.655259: 30%|### | 6/20 [00:05<00:12, 1.16it/s][I 2020-09-27 04:45:29,584] Trial 12 finished with value: 0.6590937806617987 and parameters: {'num_leaves': 112}. Best is trial 7 with value: 0.6552590163821485.
num_leaves, val_score: 0.655259: 30%|### | 6/20 [00:05<00:12, 1.16it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000922 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.264846 valid's binary_logloss: 0.665644
Early stopping, best iteration is:
[38] train's binary_logloss: 0.441545 valid's binary_logloss: 0.654899
num_leaves, val_score: 0.654899: 35%|###5 | 7/20 [00:07<00:15, 1.17s/it][I 2020-09-27 04:45:31,476] Trial 13 finished with value: 0.6548985372701535 and parameters: {'num_leaves': 256}. Best is trial 13 with value: 0.6548985372701535.
num_leaves, val_score: 0.654899: 35%|###5 | 7/20 [00:07<00:15, 1.17s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000853 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.276152 valid's binary_logloss: 0.674007
Early stopping, best iteration is:
[19] train's binary_logloss: 0.540876 valid's binary_logloss: 0.662293
num_leaves, val_score: 0.654899: 40%|#### | 8/20 [00:08<00:14, 1.17s/it][I 2020-09-27 04:45:32,652] Trial 14 finished with value: 0.6622934249190215 and parameters: {'num_leaves': 244}. Best is trial 13 with value: 0.6548985372701535.
num_leaves, val_score: 0.654899: 40%|#### | 8/20 [00:08<00:14, 1.17s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000843 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.475337 valid's binary_logloss: 0.661813
Early stopping, best iteration is:
[30] train's binary_logloss: 0.586366 valid's binary_logloss: 0.658471
num_leaves, val_score: 0.654899: 45%|####5 | 9/20 [00:08<00:11, 1.02s/it][I 2020-09-27 04:45:33,298] Trial 15 finished with value: 0.6584705405485446 and parameters: {'num_leaves': 81}. Best is trial 13 with value: 0.6548985372701535.
num_leaves, val_score: 0.654899: 45%|####5 | 9/20 [00:08<00:11, 1.02s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009872 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.32841 valid's binary_logloss: 0.676457
Early stopping, best iteration is:
[23] train's binary_logloss: 0.544414 valid's binary_logloss: 0.664223
num_leaves, val_score: 0.654899: 50%|##### | 10/20 [00:09<00:09, 1.01it/s][I 2020-09-27 04:45:34,239] Trial 16 finished with value: 0.664222940902035 and parameters: {'num_leaves': 189}. Best is trial 13 with value: 0.6548985372701535.
num_leaves, val_score: 0.654899: 50%|##### | 10/20 [00:09<00:09, 1.01it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010397 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.353294 valid's binary_logloss: 0.668112
Early stopping, best iteration is:
[38] train's binary_logloss: 0.499593 valid's binary_logloss: 0.659529
num_leaves, val_score: 0.654899: 55%|#####5 | 11/20 [00:11<00:10, 1.17s/it][I 2020-09-27 04:45:35,832] Trial 17 finished with value: 0.6595289898494625 and parameters: {'num_leaves': 166}. Best is trial 13 with value: 0.6548985372701535.
num_leaves, val_score: 0.654899: 55%|#####5 | 11/20 [00:11<00:10, 1.17s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000949 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.263628 valid's binary_logloss: 0.683347
Early stopping, best iteration is:
[22] train's binary_logloss: 0.520072 valid's binary_logloss: 0.662143
num_leaves, val_score: 0.654899: 60%|###### | 12/20 [00:12<00:09, 1.22s/it][I 2020-09-27 04:45:37,170] Trial 18 finished with value: 0.6621427253155968 and parameters: {'num_leaves': 254}. Best is trial 13 with value: 0.6548985372701535.
num_leaves, val_score: 0.654899: 60%|###### | 12/20 [00:12<00:09, 1.22s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004879 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.505464 valid's binary_logloss: 0.660159
Early stopping, best iteration is:
[34] train's binary_logloss: 0.592348 valid's binary_logloss: 0.657805
num_leaves, val_score: 0.654899: 65%|######5 | 13/20 [00:13<00:07, 1.02s/it][I 2020-09-27 04:45:37,726] Trial 19 finished with value: 0.6578051021077662 and parameters: {'num_leaves': 64}. Best is trial 13 with value: 0.6548985372701535.
num_leaves, val_score: 0.654899: 65%|######5 | 13/20 [00:13<00:07, 1.02s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000866 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.373419 valid's binary_logloss: 0.673149
Early stopping, best iteration is:
[25] train's binary_logloss: 0.55608 valid's binary_logloss: 0.664823
num_leaves, val_score: 0.654899: 70%|####### | 14/20 [00:14<00:06, 1.13s/it][I 2020-09-27 04:45:39,108] Trial 20 finished with value: 0.6648226497863418 and parameters: {'num_leaves': 150}. Best is trial 13 with value: 0.6548985372701535.
num_leaves, val_score: 0.654899: 70%|####### | 14/20 [00:14<00:06, 1.13s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000806 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.447378 valid's binary_logloss: 0.661981
Early stopping, best iteration is:
[54] train's binary_logloss: 0.521425 valid's binary_logloss: 0.658385
num_leaves, val_score: 0.654899: 75%|#######5 | 15/20 [00:15<00:05, 1.04s/it][I 2020-09-27 04:45:39,938] Trial 21 finished with value: 0.6583849556869334 and parameters: {'num_leaves': 97}. Best is trial 13 with value: 0.6548985372701535.
num_leaves, val_score: 0.654899: 75%|#######5 | 15/20 [00:15<00:05, 1.04s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004821 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650824 valid's binary_logloss: 0.661115
[200] train's binary_logloss: 0.639749 valid's binary_logloss: 0.657461
[300] train's binary_logloss: 0.632853 valid's binary_logloss: 0.656219
Early stopping, best iteration is:
[286] train's binary_logloss: 0.633724 valid's binary_logloss: 0.656027
num_leaves, val_score: 0.654899: 80%|######## | 16/20 [00:16<00:03, 1.08it/s][I 2020-09-27 04:45:40,584] Trial 22 finished with value: 0.6560272050011768 and parameters: {'num_leaves': 4}. Best is trial 13 with value: 0.6548985372701535.
num_leaves, val_score: 0.654899: 80%|######## | 16/20 [00:16<00:03, 1.08it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000847 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.295072 valid's binary_logloss: 0.672517
Early stopping, best iteration is:
[35] train's binary_logloss: 0.474148 valid's binary_logloss: 0.65871
num_leaves, val_score: 0.654899: 85%|########5 | 17/20 [00:17<00:03, 1.02s/it][I 2020-09-27 04:45:41,847] Trial 23 finished with value: 0.6587097853479928 and parameters: {'num_leaves': 222}. Best is trial 13 with value: 0.6548985372701535.
num_leaves, val_score: 0.654899: 85%|########5 | 17/20 [00:17<00:03, 1.02s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.006283 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.559059 valid's binary_logloss: 0.657941
Early stopping, best iteration is:
[78] train's binary_logloss: 0.57617 valid's binary_logloss: 0.656731
num_leaves, val_score: 0.654899: 90%|######### | 18/20 [00:18<00:02, 1.04s/it][I 2020-09-27 04:45:42,913] Trial 24 finished with value: 0.6567308128514865 and parameters: {'num_leaves': 37}. Best is trial 13 with value: 0.6548985372701535.
num_leaves, val_score: 0.654899: 90%|######### | 18/20 [00:18<00:02, 1.04s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010220 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.388417 valid's binary_logloss: 0.668977
Early stopping, best iteration is:
[38] train's binary_logloss: 0.521183 valid's binary_logloss: 0.659883
num_leaves, val_score: 0.654899: 95%|#########5| 19/20 [00:19<00:01, 1.02s/it][I 2020-09-27 04:45:43,893] Trial 25 finished with value: 0.659882543351282 and parameters: {'num_leaves': 138}. Best is trial 13 with value: 0.6548985372701535.
num_leaves, val_score: 0.654899: 95%|#########5| 19/20 [00:19<00:01, 1.02s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001162 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.465903 valid's binary_logloss: 0.661386
Early stopping, best iteration is:
[58] train's binary_logloss: 0.526905 valid's binary_logloss: 0.657041
num_leaves, val_score: 0.654899: 100%|##########| 20/20 [00:20<00:00, 1.03it/s][I 2020-09-27 04:45:44,739] Trial 26 finished with value: 0.6570410039446851 and parameters: {'num_leaves': 86}. Best is trial 13 with value: 0.6548985372701535.
num_leaves, val_score: 0.654899: 100%|##########| 20/20 [00:20<00:00, 1.01s/it]
bagging, val_score: 0.654899: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004721 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.278384 valid's binary_logloss: 0.695429
Early stopping, best iteration is:
[15] train's binary_logloss: 0.574718 valid's binary_logloss: 0.670457
bagging, val_score: 0.654899: 10%|# | 1/10 [00:01<00:10, 1.20s/it][I 2020-09-27 04:45:45,957] Trial 27 finished with value: 0.6704565961142148 and parameters: {'bagging_fraction': 0.5211974818467775, 'bagging_freq': 5}. Best is trial 27 with value: 0.6704565961142148.
bagging, val_score: 0.654899: 10%|# | 1/10 [00:01<00:10, 1.20s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.018210 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.262977 valid's binary_logloss: 0.676276
Early stopping, best iteration is:
[30] train's binary_logloss: 0.4766 valid's binary_logloss: 0.659897
bagging, val_score: 0.654899: 20%|## | 2/10 [00:02<00:10, 1.37s/it][I 2020-09-27 04:45:47,718] Trial 28 finished with value: 0.6598967542506259 and parameters: {'bagging_fraction': 0.9934815132316437, 'bagging_freq': 1}. Best is trial 28 with value: 0.6598967542506259.
bagging, val_score: 0.654899: 20%|## | 2/10 [00:02<00:10, 1.37s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000944 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.264681 valid's binary_logloss: 0.684914
Early stopping, best iteration is:
[17] train's binary_logloss: 0.549363 valid's binary_logloss: 0.665808
bagging, val_score: 0.654899: 30%|### | 3/10 [00:04<00:09, 1.35s/it][I 2020-09-27 04:45:49,025] Trial 29 finished with value: 0.6658075550967804 and parameters: {'bagging_fraction': 0.9461366644242709, 'bagging_freq': 7}. Best is trial 28 with value: 0.6598967542506259.
bagging, val_score: 0.654899: 30%|### | 3/10 [00:04<00:09, 1.35s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000798 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.281086 valid's binary_logloss: 0.707672
Early stopping, best iteration is:
[17] train's binary_logloss: 0.564424 valid's binary_logloss: 0.669487
bagging, val_score: 0.654899: 40%|#### | 4/10 [00:06<00:08, 1.49s/it][I 2020-09-27 04:45:50,825] Trial 30 finished with value: 0.6694865333360387 and parameters: {'bagging_fraction': 0.4198139622048719, 'bagging_freq': 1}. Best is trial 28 with value: 0.6598967542506259.
bagging, val_score: 0.654899: 40%|#### | 4/10 [00:06<00:08, 1.49s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000847 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.266862 valid's binary_logloss: 0.689038
Early stopping, best iteration is:
[24] train's binary_logloss: 0.514004 valid's binary_logloss: 0.66343
bagging, val_score: 0.654899: 50%|##### | 5/10 [00:07<00:07, 1.51s/it][I 2020-09-27 04:45:52,393] Trial 31 finished with value: 0.663429926936811 and parameters: {'bagging_fraction': 0.7512986978703702, 'bagging_freq': 4}. Best is trial 28 with value: 0.6598967542506259.
bagging, val_score: 0.654899: 50%|##### | 5/10 [00:07<00:07, 1.51s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004195 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.265958 valid's binary_logloss: 0.68931
Early stopping, best iteration is:
[18] train's binary_logloss: 0.549252 valid's binary_logloss: 0.6651
bagging, val_score: 0.654899: 60%|###### | 6/10 [00:09<00:05, 1.48s/it][I 2020-09-27 04:45:53,804] Trial 32 finished with value: 0.6650999977344688 and parameters: {'bagging_fraction': 0.7159980215103693, 'bagging_freq': 7}. Best is trial 28 with value: 0.6598967542506259.
bagging, val_score: 0.654899: 60%|###### | 6/10 [00:09<00:05, 1.48s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.014957 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.272788 valid's binary_logloss: 0.699964
Early stopping, best iteration is:
[22] train's binary_logloss: 0.532238 valid's binary_logloss: 0.667015
bagging, val_score: 0.654899: 70%|####### | 7/10 [00:10<00:04, 1.52s/it][I 2020-09-27 04:45:55,408] Trial 33 finished with value: 0.6670147726739027 and parameters: {'bagging_fraction': 0.5563678837629615, 'bagging_freq': 2}. Best is trial 28 with value: 0.6598967542506259.
bagging, val_score: 0.654899: 70%|####### | 7/10 [00:10<00:04, 1.52s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000817 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.267946 valid's binary_logloss: 0.681235
Early stopping, best iteration is:
[21] train's binary_logloss: 0.52815 valid's binary_logloss: 0.661059
bagging, val_score: 0.654899: 80%|######## | 8/10 [00:12<00:02, 1.47s/it][I 2020-09-27 04:45:56,765] Trial 34 finished with value: 0.6610593718244844 and parameters: {'bagging_fraction': 0.8313758530340752, 'bagging_freq': 3}. Best is trial 28 with value: 0.6598967542506259.
bagging, val_score: 0.654899: 80%|######## | 8/10 [00:12<00:02, 1.47s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001483 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.292079 valid's binary_logloss: 0.717802
Early stopping, best iteration is:
[16] train's binary_logloss: 0.575885 valid's binary_logloss: 0.667398
bagging, val_score: 0.654899: 90%|######### | 9/10 [00:13<00:01, 1.56s/it][I 2020-09-27 04:45:58,520] Trial 35 finished with value: 0.6673978773322778 and parameters: {'bagging_fraction': 0.4012246265638672, 'bagging_freq': 5}. Best is trial 28 with value: 0.6598967542506259.
bagging, val_score: 0.654899: 90%|######### | 9/10 [00:13<00:01, 1.56s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.006489 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.274203 valid's binary_logloss: 0.693572
Early stopping, best iteration is:
[23] train's binary_logloss: 0.525889 valid's binary_logloss: 0.667929
bagging, val_score: 0.654899: 100%|##########| 10/10 [00:14<00:00, 1.46s/it][I 2020-09-27 04:45:59,749] Trial 36 finished with value: 0.667928676492384 and parameters: {'bagging_fraction': 0.6154723571783186, 'bagging_freq': 6}. Best is trial 28 with value: 0.6598967542506259.
bagging, val_score: 0.654899: 100%|##########| 10/10 [00:15<00:00, 1.50s/it]
feature_fraction_stage2, val_score: 0.654899: 0%| | 0/6 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004865 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.269387 valid's binary_logloss: 0.674759
Early stopping, best iteration is:
[25] train's binary_logloss: 0.505622 valid's binary_logloss: 0.660533
feature_fraction_stage2, val_score: 0.654899: 17%|#6 | 1/6 [00:01<00:05, 1.11s/it][I 2020-09-27 04:46:00,871] Trial 37 finished with value: 0.6605329410813362 and parameters: {'feature_fraction': 0.7200000000000001}. Best is trial 37 with value: 0.6605329410813362.
feature_fraction_stage2, val_score: 0.654899: 17%|#6 | 1/6 [00:01<00:05, 1.11s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001084 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.264846 valid's binary_logloss: 0.665644
Early stopping, best iteration is:
[38] train's binary_logloss: 0.441545 valid's binary_logloss: 0.654899
feature_fraction_stage2, val_score: 0.654899: 33%|###3 | 2/6 [00:03<00:05, 1.35s/it][I 2020-09-27 04:46:02,793] Trial 38 finished with value: 0.6548985372701535 and parameters: {'feature_fraction': 0.8160000000000001}. Best is trial 38 with value: 0.6548985372701535.
feature_fraction_stage2, val_score: 0.654899: 33%|###3 | 2/6 [00:03<00:05, 1.35s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004430 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.261299 valid's binary_logloss: 0.680887
Early stopping, best iteration is:
[20] train's binary_logloss: 0.526475 valid's binary_logloss: 0.664407
feature_fraction_stage2, val_score: 0.654899: 50%|##### | 3/6 [00:04<00:03, 1.29s/it][I 2020-09-27 04:46:03,948] Trial 39 finished with value: 0.6644073193077987 and parameters: {'feature_fraction': 0.88}. Best is trial 38 with value: 0.6548985372701535.
feature_fraction_stage2, val_score: 0.654899: 50%|##### | 3/6 [00:04<00:03, 1.29s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000627 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.265461 valid's binary_logloss: 0.682897
Early stopping, best iteration is:
[24] train's binary_logloss: 0.509383 valid's binary_logloss: 0.662438
feature_fraction_stage2, val_score: 0.654899: 67%|######6 | 4/6 [00:06<00:03, 1.74s/it][I 2020-09-27 04:46:06,737] Trial 40 finished with value: 0.6624379814652182 and parameters: {'feature_fraction': 0.7520000000000001}. Best is trial 38 with value: 0.6548985372701535.
feature_fraction_stage2, val_score: 0.654899: 67%|######6 | 4/6 [00:06<00:03, 1.74s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005011 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.265461 valid's binary_logloss: 0.682897
Early stopping, best iteration is:
[24] train's binary_logloss: 0.509383 valid's binary_logloss: 0.662438
feature_fraction_stage2, val_score: 0.654899: 83%|########3 | 5/6 [00:08<00:01, 1.57s/it][I 2020-09-27 04:46:07,913] Trial 41 finished with value: 0.6624379814652182 and parameters: {'feature_fraction': 0.784}. Best is trial 38 with value: 0.6548985372701535.
feature_fraction_stage2, val_score: 0.654899: 83%|########3 | 5/6 [00:08<00:01, 1.57s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000490 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.261804 valid's binary_logloss: 0.683679
Early stopping, best iteration is:
[25] train's binary_logloss: 0.500705 valid's binary_logloss: 0.666858
feature_fraction_stage2, val_score: 0.654899: 100%|##########| 6/6 [00:09<00:00, 1.64s/it][I 2020-09-27 04:46:09,700] Trial 42 finished with value: 0.6668584743708171 and parameters: {'feature_fraction': 0.8480000000000001}. Best is trial 38 with value: 0.6548985372701535.
feature_fraction_stage2, val_score: 0.654899: 100%|##########| 6/6 [00:09<00:00, 1.66s/it]
regularization_factors, val_score: 0.654899: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001030 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.265642 valid's binary_logloss: 0.678724
Early stopping, best iteration is:
[32] train's binary_logloss: 0.469659 valid's binary_logloss: 0.660888
regularization_factors, val_score: 0.654899: 5%|5 | 1/20 [00:01<00:29, 1.58s/it][I 2020-09-27 04:46:11,296] Trial 43 finished with value: 0.6608876654466368 and parameters: {'lambda_l1': 2.395209311350816e-06, 'lambda_l2': 0.13259120292527007}. Best is trial 43 with value: 0.6608876654466368.
regularization_factors, val_score: 0.654899: 5%|5 | 1/20 [00:01<00:29, 1.58s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000545 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.390048 valid's binary_logloss: 0.672992
Early stopping, best iteration is:
[35] train's binary_logloss: 0.535479 valid's binary_logloss: 0.6619
regularization_factors, val_score: 0.654899: 10%|# | 2/20 [00:03<00:29, 1.62s/it][I 2020-09-27 04:46:13,025] Trial 44 finished with value: 0.6619000150383266 and parameters: {'lambda_l1': 6.3824374533786505, 'lambda_l2': 2.9167015559251105e-08}. Best is trial 43 with value: 0.6608876654466368.
regularization_factors, val_score: 0.654899: 10%|# | 2/20 [00:03<00:29, 1.62s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000772 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.295559 valid's binary_logloss: 0.675205
Early stopping, best iteration is:
[31] train's binary_logloss: 0.4927 valid's binary_logloss: 0.660235
regularization_factors, val_score: 0.654899: 15%|#5 | 3/20 [00:05<00:30, 1.78s/it][I 2020-09-27 04:46:15,164] Trial 45 finished with value: 0.6602349076536601 and parameters: {'lambda_l1': 1.6485218908617958, 'lambda_l2': 1.1064051625349032e-08}. Best is trial 45 with value: 0.6602349076536601.
regularization_factors, val_score: 0.654899: 15%|#5 | 3/20 [00:05<00:30, 1.78s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000792 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.263724 valid's binary_logloss: 0.666125
Early stopping, best iteration is:
[37] train's binary_logloss: 0.44514 valid's binary_logloss: 0.65487
regularization_factors, val_score: 0.654870: 20%|## | 4/20 [00:06<00:27, 1.71s/it][I 2020-09-27 04:46:16,718] Trial 46 finished with value: 0.654869889386949 and parameters: {'lambda_l1': 3.0122223559697e-08, 'lambda_l2': 0.00021523441226079608}. Best is trial 46 with value: 0.654869889386949.
regularization_factors, val_score: 0.654870: 20%|## | 4/20 [00:07<00:27, 1.71s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000873 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.263726 valid's binary_logloss: 0.666214
Early stopping, best iteration is:
[37] train's binary_logloss: 0.445141 valid's binary_logloss: 0.654998
regularization_factors, val_score: 0.654870: 25%|##5 | 5/20 [00:09<00:27, 1.80s/it][I 2020-09-27 04:46:18,736] Trial 47 finished with value: 0.6549975103511791 and parameters: {'lambda_l1': 5.742885425790403e-08, 'lambda_l2': 0.0002767008648395505}. Best is trial 46 with value: 0.654869889386949.
regularization_factors, val_score: 0.654870: 25%|##5 | 5/20 [00:09<00:27, 1.80s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001046 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.264666 valid's binary_logloss: 0.666984
Early stopping, best iteration is:
[37] train's binary_logloss: 0.445139 valid's binary_logloss: 0.654961
regularization_factors, val_score: 0.654870: 30%|### | 6/20 [00:10<00:24, 1.73s/it][I 2020-09-27 04:46:20,296] Trial 48 finished with value: 0.6549605608644221 and parameters: {'lambda_l1': 8.607940177177832e-08, 'lambda_l2': 0.0001457048379141767}. Best is trial 46 with value: 0.654869889386949.
regularization_factors, val_score: 0.654870: 30%|### | 6/20 [00:10<00:24, 1.73s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000921 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.263723 valid's binary_logloss: 0.666273
Early stopping, best iteration is:
[37] train's binary_logloss: 0.445139 valid's binary_logloss: 0.655013
regularization_factors, val_score: 0.654870: 35%|###5 | 7/20 [00:12<00:24, 1.87s/it][I 2020-09-27 04:46:22,480] Trial 49 finished with value: 0.6550133681715706 and parameters: {'lambda_l1': 1.014605269031525e-08, 'lambda_l2': 0.00017168614663450724}. Best is trial 46 with value: 0.654869889386949.
regularization_factors, val_score: 0.654870: 35%|###5 | 7/20 [00:12<00:24, 1.87s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004784 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.263723 valid's binary_logloss: 0.666224
Early stopping, best iteration is:
[37] train's binary_logloss: 0.445139 valid's binary_logloss: 0.654928
regularization_factors, val_score: 0.654870: 40%|#### | 8/20 [00:14<00:20, 1.72s/it][I 2020-09-27 04:46:23,853] Trial 50 finished with value: 0.6549280045293944 and parameters: {'lambda_l1': 1.4571400606028584e-08, 'lambda_l2': 0.00018545051962322963}. Best is trial 46 with value: 0.654869889386949.
regularization_factors, val_score: 0.654870: 40%|#### | 8/20 [00:14<00:20, 1.72s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001114 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.263723 valid's binary_logloss: 0.666238
Early stopping, best iteration is:
[37] train's binary_logloss: 0.445139 valid's binary_logloss: 0.654958
regularization_factors, val_score: 0.654870: 45%|####5 | 9/20 [00:16<00:20, 1.83s/it][I 2020-09-27 04:46:25,937] Trial 51 finished with value: 0.6549580786078956 and parameters: {'lambda_l1': 1.1439375782625129e-08, 'lambda_l2': 0.00020040952933499966}. Best is trial 46 with value: 0.654869889386949.
regularization_factors, val_score: 0.654870: 45%|####5 | 9/20 [00:16<00:20, 1.83s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004784 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.264844 valid's binary_logloss: 0.665542
Early stopping, best iteration is:
[43] train's binary_logloss: 0.421453 valid's binary_logloss: 0.654899
regularization_factors, val_score: 0.654870: 50%|##### | 10/20 [00:17<00:17, 1.71s/it][I 2020-09-27 04:46:27,383] Trial 52 finished with value: 0.6548985889877852 and parameters: {'lambda_l1': 1.3842789551680254e-08, 'lambda_l2': 5.6495755460565416e-05}. Best is trial 46 with value: 0.654869889386949.
regularization_factors, val_score: 0.654870: 50%|##### | 10/20 [00:17<00:17, 1.71s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004465 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.264846 valid's binary_logloss: 0.665605
Early stopping, best iteration is:
[38] train's binary_logloss: 0.441545 valid's binary_logloss: 0.654867
regularization_factors, val_score: 0.654867: 55%|#####5 | 11/20 [00:19<00:16, 1.80s/it][I 2020-09-27 04:46:29,389] Trial 53 finished with value: 0.6548668432849638 and parameters: {'lambda_l1': 1.189642528157414e-08, 'lambda_l2': 5.203275530782603e-06}. Best is trial 53 with value: 0.6548668432849638.
regularization_factors, val_score: 0.654867: 55%|#####5 | 11/20 [00:19<00:16, 1.80s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000851 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.264846 valid's binary_logloss: 0.665516
Early stopping, best iteration is:
[38] train's binary_logloss: 0.441545 valid's binary_logloss: 0.654838
regularization_factors, val_score: 0.654838: 60%|###### | 12/20 [00:21<00:13, 1.73s/it][I 2020-09-27 04:46:30,946] Trial 54 finished with value: 0.6548376800578837 and parameters: {'lambda_l1': 2.1670289548196517e-06, 'lambda_l2': 1.661838347245463e-06}. Best is trial 54 with value: 0.6548376800578837.
regularization_factors, val_score: 0.654838: 60%|###### | 12/20 [00:21<00:13, 1.73s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000858 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.264846 valid's binary_logloss: 0.66563
Early stopping, best iteration is:
[38] train's binary_logloss: 0.441545 valid's binary_logloss: 0.654905
regularization_factors, val_score: 0.654838: 65%|######5 | 13/20 [00:23<00:12, 1.84s/it][I 2020-09-27 04:46:33,033] Trial 55 finished with value: 0.6549047704174789 and parameters: {'lambda_l1': 1.3047485080865155e-05, 'lambda_l2': 1.1388372443834088e-06}. Best is trial 54 with value: 0.6548376800578837.
regularization_factors, val_score: 0.654838: 65%|######5 | 13/20 [00:23<00:12, 1.84s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000859 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.264846 valid's binary_logloss: 0.665482
Early stopping, best iteration is:
[38] train's binary_logloss: 0.441545 valid's binary_logloss: 0.654874
regularization_factors, val_score: 0.654838: 70%|####### | 14/20 [00:24<00:10, 1.75s/it][I 2020-09-27 04:46:34,599] Trial 56 finished with value: 0.6548744228748046 and parameters: {'lambda_l1': 1.677207413460934e-06, 'lambda_l2': 2.3547411212794384e-06}. Best is trial 54 with value: 0.6548376800578837.
regularization_factors, val_score: 0.654838: 70%|####### | 14/20 [00:24<00:10, 1.75s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000869 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.264846 valid's binary_logloss: 0.665557
Early stopping, best iteration is:
[38] train's binary_logloss: 0.441545 valid's binary_logloss: 0.654891
regularization_factors, val_score: 0.654838: 75%|#######5 | 15/20 [00:27<00:09, 1.87s/it][I 2020-09-27 04:46:36,725] Trial 57 finished with value: 0.65489106531928 and parameters: {'lambda_l1': 2.808990582697963e-06, 'lambda_l2': 3.680142856828128e-06}. Best is trial 54 with value: 0.6548376800578837.
regularization_factors, val_score: 0.654838: 75%|#######5 | 15/20 [00:27<00:09, 1.87s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005001 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.264846 valid's binary_logloss: 0.665537
Early stopping, best iteration is:
[38] train's binary_logloss: 0.441545 valid's binary_logloss: 0.654883
regularization_factors, val_score: 0.654838: 80%|######## | 16/20 [00:28<00:07, 1.75s/it][I 2020-09-27 04:46:38,214] Trial 58 finished with value: 0.6548833147276146 and parameters: {'lambda_l1': 3.548471771976571e-06, 'lambda_l2': 1.3069611448432158e-06}. Best is trial 54 with value: 0.6548376800578837.
regularization_factors, val_score: 0.654838: 80%|######## | 16/20 [00:28<00:07, 1.75s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000875 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.264846 valid's binary_logloss: 0.665518
Early stopping, best iteration is:
[38] train's binary_logloss: 0.441545 valid's binary_logloss: 0.654891
regularization_factors, val_score: 0.654838: 85%|########5 | 17/20 [00:30<00:05, 1.86s/it][I 2020-09-27 04:46:40,318] Trial 59 finished with value: 0.654890562836266 and parameters: {'lambda_l1': 3.6573101362801625e-06, 'lambda_l2': 1.3655499334616953e-06}. Best is trial 54 with value: 0.6548376800578837.
regularization_factors, val_score: 0.654838: 85%|########5 | 17/20 [00:30<00:05, 1.86s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000836 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.263781 valid's binary_logloss: 0.676626
Early stopping, best iteration is:
[41] train's binary_logloss: 0.42877 valid's binary_logloss: 0.654741
regularization_factors, val_score: 0.654741: 90%|######### | 18/20 [00:32<00:03, 1.78s/it][I 2020-09-27 04:46:41,916] Trial 60 finished with value: 0.6547406780914222 and parameters: {'lambda_l1': 8.237457561826076e-05, 'lambda_l2': 7.033731015507948e-07}. Best is trial 60 with value: 0.6547406780914222.
regularization_factors, val_score: 0.654741: 90%|######### | 18/20 [00:32<00:03, 1.78s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000852 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.264845 valid's binary_logloss: 0.66557
Early stopping, best iteration is:
[38] train's binary_logloss: 0.441546 valid's binary_logloss: 0.654905
regularization_factors, val_score: 0.654741: 95%|#########5| 19/20 [00:34<00:01, 1.87s/it][I 2020-09-27 04:46:43,993] Trial 61 finished with value: 0.6549054519516266 and parameters: {'lambda_l1': 8.016334404835216e-05, 'lambda_l2': 8.515948497846046e-07}. Best is trial 60 with value: 0.6547406780914222.
regularization_factors, val_score: 0.654741: 95%|#########5| 19/20 [00:34<00:01, 1.87s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000852 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.264846 valid's binary_logloss: 0.665533
Early stopping, best iteration is:
[38] train's binary_logloss: 0.441545 valid's binary_logloss: 0.65488
regularization_factors, val_score: 0.654741: 100%|##########| 20/20 [00:35<00:00, 1.78s/it][I 2020-09-27 04:46:45,569] Trial 62 finished with value: 0.6548801766746073 and parameters: {'lambda_l1': 6.807709893234408e-07, 'lambda_l2': 1.3368946882298463e-06}. Best is trial 60 with value: 0.6547406780914222.
regularization_factors, val_score: 0.654741: 100%|##########| 20/20 [00:35<00:00, 1.79s/it]
min_data_in_leaf, val_score: 0.654741: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000788 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.26674 valid's binary_logloss: 0.673371
Early stopping, best iteration is:
[37] train's binary_logloss: 0.44854 valid's binary_logloss: 0.659997
min_data_in_leaf, val_score: 0.654741: 20%|## | 1/5 [00:02<00:08, 2.00s/it][I 2020-09-27 04:46:47,591] Trial 63 finished with value: 0.6599970673875724 and parameters: {'min_child_samples': 25}. Best is trial 63 with value: 0.6599970673875724.
min_data_in_leaf, val_score: 0.654741: 20%|## | 1/5 [00:02<00:08, 2.00s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000851 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.259932 valid's binary_logloss: 0.674681
Early stopping, best iteration is:
[27] train's binary_logloss: 0.485136 valid's binary_logloss: 0.663118
min_data_in_leaf, val_score: 0.654741: 40%|#### | 2/5 [00:03<00:05, 1.90s/it][I 2020-09-27 04:46:49,234] Trial 64 finished with value: 0.6631181004712279 and parameters: {'min_child_samples': 5}. Best is trial 63 with value: 0.6599970673875724.
min_data_in_leaf, val_score: 0.654741: 40%|#### | 2/5 [00:03<00:05, 1.90s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000813 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.277711 valid's binary_logloss: 0.677648
Early stopping, best iteration is:
[35] train's binary_logloss: 0.469589 valid's binary_logloss: 0.66334
min_data_in_leaf, val_score: 0.654741: 60%|###### | 3/5 [00:05<00:03, 1.93s/it][I 2020-09-27 04:46:51,228] Trial 65 finished with value: 0.6633400797460791 and parameters: {'min_child_samples': 50}. Best is trial 63 with value: 0.6599970673875724.
min_data_in_leaf, val_score: 0.654741: 60%|###### | 3/5 [00:05<00:03, 1.93s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000892 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.262769 valid's binary_logloss: 0.674802
Early stopping, best iteration is:
[35] train's binary_logloss: 0.44869 valid's binary_logloss: 0.662608
min_data_in_leaf, val_score: 0.654741: 80%|######## | 4/5 [00:07<00:01, 1.86s/it][I 2020-09-27 04:46:52,928] Trial 66 finished with value: 0.6626078522269558 and parameters: {'min_child_samples': 10}. Best is trial 63 with value: 0.6599970673875724.
min_data_in_leaf, val_score: 0.654741: 80%|######## | 4/5 [00:07<00:01, 1.86s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000877 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4238
[LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156
[LightGBM] [Info] Start training from score -0.023156
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
Training until validation scores don't improve for 100 rounds
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[100] train's binary_logloss: 0.355756 valid's binary_logloss: 0.678018
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
[LightGBM] [Warning] No further splits with positive gain, best gain: -inf
Early stopping, best iteration is:
[32] train's binary_logloss: 0.530938 valid's binary_logloss: 0.659107
min_data_in_leaf, val_score: 0.654741: 100%|##########| 5/5 [00:10<00:00, 2.23s/it][I 2020-09-27 04:46:56,042] Trial 67 finished with value: 0.6591070735201429 and parameters: {'min_child_samples': 100}. Best is trial 67 with value: 0.6591070735201429.
min_data_in_leaf, val_score: 0.654741: 100%|##########| 5/5 [00:10<00:00, 2.09s/it]
Fold : 8
[I 2020-09-27 04:46:56,086] A new study created in memory with name: no-name-0ac6def8-9fc0-44ca-a007-4dfc936cce3b
feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000362 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.582362 valid's binary_logloss: 0.657172
Early stopping, best iteration is:
[75] train's binary_logloss: 0.598686 valid's binary_logloss: 0.656832
feature_fraction, val_score: 0.656832: 14%|#4 | 1/7 [00:00<00:03, 1.84it/s][I 2020-09-27 04:46:56,647] Trial 0 finished with value: 0.6568322981471987 and parameters: {'feature_fraction': 0.4}. Best is trial 0 with value: 0.6568322981471987.
feature_fraction, val_score: 0.656832: 14%|#4 | 1/7 [00:00<00:03, 1.84it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004661 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.575682 valid's binary_logloss: 0.657591
Early stopping, best iteration is:
[65] train's binary_logloss: 0.598895 valid's binary_logloss: 0.65488
feature_fraction, val_score: 0.654880: 29%|##8 | 2/7 [00:01<00:02, 1.90it/s][I 2020-09-27 04:46:57,132] Trial 1 finished with value: 0.654879918185127 and parameters: {'feature_fraction': 0.7}. Best is trial 1 with value: 0.654879918185127.
feature_fraction, val_score: 0.654880: 29%|##8 | 2/7 [00:01<00:02, 1.90it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000886 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.573899 valid's binary_logloss: 0.657206
Early stopping, best iteration is:
[70] train's binary_logloss: 0.593743 valid's binary_logloss: 0.655846
feature_fraction, val_score: 0.654880: 43%|####2 | 3/7 [00:01<00:02, 1.85it/s][I 2020-09-27 04:46:57,704] Trial 2 finished with value: 0.6558455623356558 and parameters: {'feature_fraction': 0.8}. Best is trial 1 with value: 0.654879918185127.
feature_fraction, val_score: 0.654880: 43%|####2 | 3/7 [00:01<00:02, 1.85it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000972 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.57084 valid's binary_logloss: 0.654546
[200] train's binary_logloss: 0.513329 valid's binary_logloss: 0.658828
Early stopping, best iteration is:
[113] train's binary_logloss: 0.562928 valid's binary_logloss: 0.654073
feature_fraction, val_score: 0.654073: 57%|#####7 | 4/7 [00:02<00:01, 1.73it/s][I 2020-09-27 04:46:58,369] Trial 3 finished with value: 0.6540730143979465 and parameters: {'feature_fraction': 1.0}. Best is trial 3 with value: 0.6540730143979465.
feature_fraction, val_score: 0.654073: 57%|#####7 | 4/7 [00:02<00:01, 1.73it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004601 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.579014 valid's binary_logloss: 0.655596
Early stopping, best iteration is:
[59] train's binary_logloss: 0.606682 valid's binary_logloss: 0.6549
feature_fraction, val_score: 0.654073: 71%|#######1 | 5/7 [00:02<00:01, 1.63it/s][I 2020-09-27 04:46:59,068] Trial 4 finished with value: 0.6548998752828071 and parameters: {'feature_fraction': 0.5}. Best is trial 3 with value: 0.6540730143979465.
feature_fraction, val_score: 0.654073: 71%|#######1 | 5/7 [00:02<00:01, 1.63it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011929 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.572184 valid's binary_logloss: 0.658179
Early stopping, best iteration is:
[67] train's binary_logloss: 0.594979 valid's binary_logloss: 0.656087
feature_fraction, val_score: 0.654073: 86%|########5 | 6/7 [00:03<00:00, 1.52it/s][I 2020-09-27 04:46:59,834] Trial 5 finished with value: 0.6560873120955287 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 3 with value: 0.6540730143979465.
feature_fraction, val_score: 0.654073: 86%|########5 | 6/7 [00:03<00:00, 1.52it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004851 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.576475 valid's binary_logloss: 0.652717
Early stopping, best iteration is:
[82] train's binary_logloss: 0.587886 valid's binary_logloss: 0.65223
feature_fraction, val_score: 0.652230: 100%|##########| 7/7 [00:04<00:00, 1.64it/s][I 2020-09-27 04:47:00,328] Trial 6 finished with value: 0.6522296496116577 and parameters: {'feature_fraction': 0.6}. Best is trial 6 with value: 0.6522296496116577.
feature_fraction, val_score: 0.652230: 100%|##########| 7/7 [00:04<00:00, 1.65it/s]
num_leaves, val_score: 0.652230: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004842 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.326219 valid's binary_logloss: 0.670645
Early stopping, best iteration is:
[35] train's binary_logloss: 0.497287 valid's binary_logloss: 0.658257
num_leaves, val_score: 0.652230: 5%|5 | 1/20 [00:00<00:17, 1.08it/s][I 2020-09-27 04:47:01,277] Trial 7 finished with value: 0.6582573078988242 and parameters: {'num_leaves': 200}. Best is trial 7 with value: 0.6582573078988242.
num_leaves, val_score: 0.652230: 5%|5 | 1/20 [00:00<00:17, 1.08it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004568 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.281259 valid's binary_logloss: 0.679658
Early stopping, best iteration is:
[36] train's binary_logloss: 0.462423 valid's binary_logloss: 0.663543
num_leaves, val_score: 0.652230: 10%|# | 2/20 [00:02<00:17, 1.03it/s][I 2020-09-27 04:47:02,352] Trial 8 finished with value: 0.6635430249480666 and parameters: {'num_leaves': 251}. Best is trial 7 with value: 0.6582573078988242.
num_leaves, val_score: 0.652230: 10%|# | 2/20 [00:02<00:17, 1.03it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005025 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657954 valid's binary_logloss: 0.664118
[200] train's binary_logloss: 0.647761 valid's binary_logloss: 0.655763
[300] train's binary_logloss: 0.642622 valid's binary_logloss: 0.652688
[400] train's binary_logloss: 0.638865 valid's binary_logloss: 0.651406
[500] train's binary_logloss: 0.635649 valid's binary_logloss: 0.650648
[600] train's binary_logloss: 0.632824 valid's binary_logloss: 0.650562
[700] train's binary_logloss: 0.630241 valid's binary_logloss: 0.650393
Early stopping, best iteration is:
[694] train's binary_logloss: 0.630392 valid's binary_logloss: 0.650237
num_leaves, val_score: 0.650237: 15%|#5 | 3/20 [00:03<00:19, 1.17s/it][I 2020-09-27 04:47:03,969] Trial 9 finished with value: 0.65023693049223 and parameters: {'num_leaves': 3}. Best is trial 9 with value: 0.65023693049223.
num_leaves, val_score: 0.650237: 15%|#5 | 3/20 [00:03<00:19, 1.17s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004669 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657954 valid's binary_logloss: 0.664118
[200] train's binary_logloss: 0.647761 valid's binary_logloss: 0.655763
[300] train's binary_logloss: 0.642622 valid's binary_logloss: 0.652688
[400] train's binary_logloss: 0.638865 valid's binary_logloss: 0.651406
[500] train's binary_logloss: 0.635649 valid's binary_logloss: 0.650648
[600] train's binary_logloss: 0.632824 valid's binary_logloss: 0.650562
[700] train's binary_logloss: 0.630241 valid's binary_logloss: 0.650393
Early stopping, best iteration is:
[694] train's binary_logloss: 0.630392 valid's binary_logloss: 0.650237
num_leaves, val_score: 0.650237: 20%|## | 4/20 [00:04<00:18, 1.18s/it][I 2020-09-27 04:47:05,175] Trial 10 finished with value: 0.6502369304922299 and parameters: {'num_leaves': 3}. Best is trial 10 with value: 0.6502369304922299.
num_leaves, val_score: 0.650237: 20%|## | 4/20 [00:04<00:18, 1.18s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005031 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.628976 valid's binary_logloss: 0.654294
[200] train's binary_logloss: 0.608343 valid's binary_logloss: 0.65336
Early stopping, best iteration is:
[176] train's binary_logloss: 0.613043 valid's binary_logloss: 0.652178
num_leaves, val_score: 0.650237: 25%|##5 | 5/20 [00:05<00:14, 1.01it/s][I 2020-09-27 04:47:05,732] Trial 11 finished with value: 0.6521775910881051 and parameters: {'num_leaves': 10}. Best is trial 10 with value: 0.6502369304922299.
num_leaves, val_score: 0.650237: 25%|##5 | 5/20 [00:05<00:14, 1.01it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004975 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.635408 valid's binary_logloss: 0.65585
[200] train's binary_logloss: 0.618644 valid's binary_logloss: 0.654948
Early stopping, best iteration is:
[161] train's binary_logloss: 0.624573 valid's binary_logloss: 0.6542
num_leaves, val_score: 0.650237: 30%|### | 6/20 [00:05<00:11, 1.17it/s][I 2020-09-27 04:47:06,273] Trial 12 finished with value: 0.6541997386028066 and parameters: {'num_leaves': 8}. Best is trial 10 with value: 0.6502369304922299.
num_leaves, val_score: 0.650237: 30%|### | 6/20 [00:05<00:11, 1.17it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000898 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.488946 valid's binary_logloss: 0.659659
Early stopping, best iteration is:
[39] train's binary_logloss: 0.5773 valid's binary_logloss: 0.65642
num_leaves, val_score: 0.650237: 35%|###5 | 7/20 [00:07<00:12, 1.04it/s][I 2020-09-27 04:47:07,468] Trial 13 finished with value: 0.6564196118293095 and parameters: {'num_leaves': 76}. Best is trial 10 with value: 0.6502369304922299.
num_leaves, val_score: 0.650237: 35%|###5 | 7/20 [00:07<00:12, 1.04it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000884 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.498154 valid's binary_logloss: 0.662208
Early stopping, best iteration is:
[39] train's binary_logloss: 0.581162 valid's binary_logloss: 0.657338
num_leaves, val_score: 0.650237: 40%|#### | 8/20 [00:07<00:10, 1.16it/s][I 2020-09-27 04:47:08,100] Trial 14 finished with value: 0.6573383283023841 and parameters: {'num_leaves': 71}. Best is trial 10 with value: 0.6502369304922299.
num_leaves, val_score: 0.650237: 40%|#### | 8/20 [00:07<00:10, 1.16it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004661 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.646593 valid's binary_logloss: 0.657131
[200] train's binary_logloss: 0.634403 valid's binary_logloss: 0.652331
[300] train's binary_logloss: 0.625932 valid's binary_logloss: 0.651212
Early stopping, best iteration is:
[289] train's binary_logloss: 0.626825 valid's binary_logloss: 0.650977
num_leaves, val_score: 0.650237: 45%|####5 | 9/20 [00:08<00:08, 1.27it/s][I 2020-09-27 04:47:08,723] Trial 15 finished with value: 0.6509770662129648 and parameters: {'num_leaves': 5}. Best is trial 10 with value: 0.6502369304922299.
num_leaves, val_score: 0.650237: 45%|####5 | 9/20 [00:08<00:08, 1.27it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004588 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.498154 valid's binary_logloss: 0.662208
Early stopping, best iteration is:
[39] train's binary_logloss: 0.581162 valid's binary_logloss: 0.657338
num_leaves, val_score: 0.650237: 50%|##### | 10/20 [00:08<00:07, 1.40it/s][I 2020-09-27 04:47:09,256] Trial 16 finished with value: 0.6573383283023841 and parameters: {'num_leaves': 71}. Best is trial 10 with value: 0.6502369304922299.
num_leaves, val_score: 0.650237: 50%|##### | 10/20 [00:08<00:07, 1.40it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012770 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.392369 valid's binary_logloss: 0.669429
Early stopping, best iteration is:
[45] train's binary_logloss: 0.507218 valid's binary_logloss: 0.65932
num_leaves, val_score: 0.650237: 55%|#####5 | 11/20 [00:09<00:06, 1.35it/s][I 2020-09-27 04:47:10,058] Trial 17 finished with value: 0.659319879723429 and parameters: {'num_leaves': 141}. Best is trial 10 with value: 0.6502369304922299.
num_leaves, val_score: 0.650237: 55%|#####5 | 11/20 [00:09<00:06, 1.35it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010521 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.553596 valid's binary_logloss: 0.658319
Early stopping, best iteration is:
[54] train's binary_logloss: 0.593584 valid's binary_logloss: 0.656096
num_leaves, val_score: 0.650237: 60%|###### | 12/20 [00:10<00:06, 1.25it/s][I 2020-09-27 04:47:10,992] Trial 18 finished with value: 0.6560958307756177 and parameters: {'num_leaves': 42}. Best is trial 10 with value: 0.6502369304922299.
num_leaves, val_score: 0.650237: 60%|###### | 12/20 [00:10<00:06, 1.25it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001062 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.409616 valid's binary_logloss: 0.665328
Early stopping, best iteration is:
[37] train's binary_logloss: 0.53918 valid's binary_logloss: 0.658274
num_leaves, val_score: 0.650237: 65%|######5 | 13/20 [00:11<00:05, 1.19it/s][I 2020-09-27 04:47:11,928] Trial 19 finished with value: 0.6582743650249848 and parameters: {'num_leaves': 128}. Best is trial 10 with value: 0.6502369304922299.
num_leaves, val_score: 0.650237: 65%|######5 | 13/20 [00:11<00:05, 1.19it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004694 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.563913 valid's binary_logloss: 0.657197
Early stopping, best iteration is:
[77] train's binary_logloss: 0.580794 valid's binary_logloss: 0.654988
num_leaves, val_score: 0.650237: 70%|####### | 14/20 [00:12<00:04, 1.37it/s][I 2020-09-27 04:47:12,412] Trial 20 finished with value: 0.6549884052454669 and parameters: {'num_leaves': 37}. Best is trial 10 with value: 0.6502369304922299.
num_leaves, val_score: 0.650237: 70%|####### | 14/20 [00:12<00:04, 1.37it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009279 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657954 valid's binary_logloss: 0.664118
[200] train's binary_logloss: 0.647761 valid's binary_logloss: 0.655763
[300] train's binary_logloss: 0.642622 valid's binary_logloss: 0.652688
[400] train's binary_logloss: 0.638865 valid's binary_logloss: 0.651406
[500] train's binary_logloss: 0.635649 valid's binary_logloss: 0.650648
[600] train's binary_logloss: 0.632824 valid's binary_logloss: 0.650562
[700] train's binary_logloss: 0.630241 valid's binary_logloss: 0.650393
Early stopping, best iteration is:
[694] train's binary_logloss: 0.630392 valid's binary_logloss: 0.650237
num_leaves, val_score: 0.650237: 75%|#######5 | 15/20 [00:13<00:04, 1.18it/s][I 2020-09-27 04:47:13,534] Trial 21 finished with value: 0.6502369304922299 and parameters: {'num_leaves': 3}. Best is trial 10 with value: 0.6502369304922299.
num_leaves, val_score: 0.650237: 75%|#######5 | 15/20 [00:13<00:04, 1.18it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000785 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.646593 valid's binary_logloss: 0.657131
[200] train's binary_logloss: 0.634403 valid's binary_logloss: 0.652331
[300] train's binary_logloss: 0.625932 valid's binary_logloss: 0.651212
Early stopping, best iteration is:
[289] train's binary_logloss: 0.626825 valid's binary_logloss: 0.650977
num_leaves, val_score: 0.650237: 80%|######## | 16/20 [00:14<00:03, 1.08it/s][I 2020-09-27 04:47:14,641] Trial 22 finished with value: 0.6509770662129648 and parameters: {'num_leaves': 5}. Best is trial 10 with value: 0.6502369304922299.
num_leaves, val_score: 0.650237: 80%|######## | 16/20 [00:14<00:03, 1.08it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.019722 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.565177 valid's binary_logloss: 0.655115
Early stopping, best iteration is:
[72] train's binary_logloss: 0.585373 valid's binary_logloss: 0.65339
num_leaves, val_score: 0.650237: 85%|########5 | 17/20 [00:14<00:02, 1.20it/s][I 2020-09-27 04:47:15,254] Trial 23 finished with value: 0.6533900763638031 and parameters: {'num_leaves': 36}. Best is trial 10 with value: 0.6502369304922299.
num_leaves, val_score: 0.650237: 85%|########5 | 17/20 [00:14<00:02, 1.20it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004882 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.565177 valid's binary_logloss: 0.655115
Early stopping, best iteration is:
[72] train's binary_logloss: 0.585373 valid's binary_logloss: 0.65339
num_leaves, val_score: 0.650237: 90%|######### | 18/20 [00:15<00:01, 1.36it/s][I 2020-09-27 04:47:15,759] Trial 24 finished with value: 0.6533900763638031 and parameters: {'num_leaves': 36}. Best is trial 10 with value: 0.6502369304922299.
num_leaves, val_score: 0.650237: 90%|######### | 18/20 [00:15<00:01, 1.36it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009941 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.668764 valid's binary_logloss: 0.672745
[200] train's binary_logloss: 0.65923 valid's binary_logloss: 0.663108
[300] train's binary_logloss: 0.654067 valid's binary_logloss: 0.658349
[400] train's binary_logloss: 0.651004 valid's binary_logloss: 0.655269
[500] train's binary_logloss: 0.649032 valid's binary_logloss: 0.653548
[600] train's binary_logloss: 0.647704 valid's binary_logloss: 0.652575
[700] train's binary_logloss: 0.646761 valid's binary_logloss: 0.651831
[800] train's binary_logloss: 0.646051 valid's binary_logloss: 0.651481
[900] train's binary_logloss: 0.645487 valid's binary_logloss: 0.651115
[1000] train's binary_logloss: 0.645019 valid's binary_logloss: 0.651086
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.645019 valid's binary_logloss: 0.651086
num_leaves, val_score: 0.650237: 95%|#########5| 19/20 [00:16<00:00, 1.12it/s][I 2020-09-27 04:47:17,024] Trial 25 finished with value: 0.6510861135332809 and parameters: {'num_leaves': 2}. Best is trial 10 with value: 0.6502369304922299.
num_leaves, val_score: 0.650237: 95%|#########5| 19/20 [00:16<00:00, 1.12it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005016 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.424526 valid's binary_logloss: 0.659895
Early stopping, best iteration is:
[33] train's binary_logloss: 0.556295 valid's binary_logloss: 0.654704
num_leaves, val_score: 0.650237: 100%|##########| 20/20 [00:17<00:00, 1.20it/s][I 2020-09-27 04:47:17,712] Trial 26 finished with value: 0.6547044966874118 and parameters: {'num_leaves': 119}. Best is trial 10 with value: 0.6502369304922299.
num_leaves, val_score: 0.650237: 100%|##########| 20/20 [00:17<00:00, 1.15it/s]
bagging, val_score: 0.650237: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004738 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.658045 valid's binary_logloss: 0.664127
[200] train's binary_logloss: 0.647774 valid's binary_logloss: 0.655872
[300] train's binary_logloss: 0.642497 valid's binary_logloss: 0.65234
[400] train's binary_logloss: 0.638669 valid's binary_logloss: 0.651347
[500] train's binary_logloss: 0.635478 valid's binary_logloss: 0.651168
Early stopping, best iteration is:
[453] train's binary_logloss: 0.63696 valid's binary_logloss: 0.650827
bagging, val_score: 0.650237: 10%|# | 1/10 [00:01<00:12, 1.44s/it][I 2020-09-27 04:47:19,168] Trial 27 finished with value: 0.6508272913841253 and parameters: {'bagging_fraction': 0.9963833291336945, 'bagging_freq': 6}. Best is trial 27 with value: 0.6508272913841253.
bagging, val_score: 0.650237: 10%|# | 1/10 [00:01<00:12, 1.44s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004596 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.655889 valid's binary_logloss: 0.659928
[200] train's binary_logloss: 0.646169 valid's binary_logloss: 0.654108
[300] train's binary_logloss: 0.641381 valid's binary_logloss: 0.651946
[400] train's binary_logloss: 0.637836 valid's binary_logloss: 0.650942
Early stopping, best iteration is:
[389] train's binary_logloss: 0.638252 valid's binary_logloss: 0.650231
bagging, val_score: 0.650231: 20%|## | 2/10 [00:02<00:10, 1.27s/it][I 2020-09-27 04:47:20,036] Trial 28 finished with value: 0.6502306281879949 and parameters: {'bagging_fraction': 0.4514619063105707, 'bagging_freq': 1}. Best is trial 28 with value: 0.6502306281879949.
bagging, val_score: 0.650231: 20%|## | 2/10 [00:02<00:10, 1.27s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004540 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.655825 valid's binary_logloss: 0.660152
[200] train's binary_logloss: 0.646253 valid's binary_logloss: 0.653794
[300] train's binary_logloss: 0.64168 valid's binary_logloss: 0.652426
[400] train's binary_logloss: 0.638103 valid's binary_logloss: 0.651436
Early stopping, best iteration is:
[381] train's binary_logloss: 0.63871 valid's binary_logloss: 0.650844
bagging, val_score: 0.650231: 30%|### | 3/10 [00:03<00:07, 1.13s/it][I 2020-09-27 04:47:20,849] Trial 29 finished with value: 0.6508442537589473 and parameters: {'bagging_fraction': 0.4102073189555854, 'bagging_freq': 1}. Best is trial 28 with value: 0.6502306281879949.
bagging, val_score: 0.650231: 30%|### | 3/10 [00:03<00:07, 1.13s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004944 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.655879 valid's binary_logloss: 0.660227
[200] train's binary_logloss: 0.646389 valid's binary_logloss: 0.653638
[300] train's binary_logloss: 0.641982 valid's binary_logloss: 0.652522
[400] train's binary_logloss: 0.638289 valid's binary_logloss: 0.652467
Early stopping, best iteration is:
[387] train's binary_logloss: 0.63872 valid's binary_logloss: 0.652011
bagging, val_score: 0.650231: 40%|#### | 4/10 [00:03<00:06, 1.04s/it][I 2020-09-27 04:47:21,683] Trial 30 finished with value: 0.6520113408937251 and parameters: {'bagging_fraction': 0.4058702957626291, 'bagging_freq': 1}. Best is trial 28 with value: 0.6502306281879949.
bagging, val_score: 0.650231: 40%|#### | 4/10 [00:03<00:06, 1.04s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008559 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656756 valid's binary_logloss: 0.662314
[200] train's binary_logloss: 0.646612 valid's binary_logloss: 0.653269
[300] train's binary_logloss: 0.641725 valid's binary_logloss: 0.650745
[400] train's binary_logloss: 0.638127 valid's binary_logloss: 0.650437
Early stopping, best iteration is:
[386] train's binary_logloss: 0.638596 valid's binary_logloss: 0.649682
bagging, val_score: 0.649682: 50%|##### | 5/10 [00:05<00:05, 1.10s/it][I 2020-09-27 04:47:22,921] Trial 31 finished with value: 0.6496818435732342 and parameters: {'bagging_fraction': 0.6494590447691004, 'bagging_freq': 3}. Best is trial 31 with value: 0.6496818435732342.
bagging, val_score: 0.649682: 50%|##### | 5/10 [00:05<00:05, 1.10s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004698 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.65681 valid's binary_logloss: 0.661741
[200] train's binary_logloss: 0.646579 valid's binary_logloss: 0.653526
[300] train's binary_logloss: 0.641414 valid's binary_logloss: 0.651164
[400] train's binary_logloss: 0.637939 valid's binary_logloss: 0.650273
Early stopping, best iteration is:
[386] train's binary_logloss: 0.638417 valid's binary_logloss: 0.649953
bagging, val_score: 0.649682: 60%|###### | 6/10 [00:05<00:03, 1.00it/s][I 2020-09-27 04:47:23,670] Trial 32 finished with value: 0.6499526525656477 and parameters: {'bagging_fraction': 0.6912299232252367, 'bagging_freq': 3}. Best is trial 31 with value: 0.6496818435732342.
bagging, val_score: 0.649682: 60%|###### | 6/10 [00:05<00:03, 1.00it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004577 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656917 valid's binary_logloss: 0.662005
[200] train's binary_logloss: 0.646694 valid's binary_logloss: 0.653638
[300] train's binary_logloss: 0.641556 valid's binary_logloss: 0.651449
[400] train's binary_logloss: 0.637865 valid's binary_logloss: 0.650759
Early stopping, best iteration is:
[387] train's binary_logloss: 0.638305 valid's binary_logloss: 0.650278
bagging, val_score: 0.649682: 70%|####### | 7/10 [00:06<00:02, 1.09it/s][I 2020-09-27 04:47:24,413] Trial 33 finished with value: 0.6502781308682245 and parameters: {'bagging_fraction': 0.7006917452683191, 'bagging_freq': 3}. Best is trial 31 with value: 0.6496818435732342.
bagging, val_score: 0.649682: 70%|####### | 7/10 [00:06<00:02, 1.09it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005647 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656936 valid's binary_logloss: 0.66276
[200] train's binary_logloss: 0.646675 valid's binary_logloss: 0.653779
[300] train's binary_logloss: 0.641779 valid's binary_logloss: 0.651157
[400] train's binary_logloss: 0.638073 valid's binary_logloss: 0.650615
Early stopping, best iteration is:
[385] train's binary_logloss: 0.638614 valid's binary_logloss: 0.649988
bagging, val_score: 0.649682: 80%|######## | 8/10 [00:07<00:01, 1.14it/s][I 2020-09-27 04:47:25,180] Trial 34 finished with value: 0.6499876467524935 and parameters: {'bagging_fraction': 0.6697359412743568, 'bagging_freq': 3}. Best is trial 31 with value: 0.6496818435732342.
bagging, val_score: 0.649682: 80%|######## | 8/10 [00:07<00:01, 1.14it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005015 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656928 valid's binary_logloss: 0.661654
[200] train's binary_logloss: 0.646722 valid's binary_logloss: 0.652579
[300] train's binary_logloss: 0.641759 valid's binary_logloss: 0.65045
[400] train's binary_logloss: 0.638172 valid's binary_logloss: 0.649763
[500] train's binary_logloss: 0.634789 valid's binary_logloss: 0.649596
Early stopping, best iteration is:
[472] train's binary_logloss: 0.635712 valid's binary_logloss: 0.649063
bagging, val_score: 0.649063: 90%|######### | 9/10 [00:08<00:01, 1.00s/it][I 2020-09-27 04:47:26,487] Trial 35 finished with value: 0.649063213862809 and parameters: {'bagging_fraction': 0.6949061090676492, 'bagging_freq': 3}. Best is trial 35 with value: 0.649063213862809.
bagging, val_score: 0.649063: 90%|######### | 9/10 [00:08<00:01, 1.00s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.006251 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656842 valid's binary_logloss: 0.662647
[200] train's binary_logloss: 0.646752 valid's binary_logloss: 0.654227
[300] train's binary_logloss: 0.641722 valid's binary_logloss: 0.65187
[400] train's binary_logloss: 0.638055 valid's binary_logloss: 0.651604
[500] train's binary_logloss: 0.634682 valid's binary_logloss: 0.651328
Early stopping, best iteration is:
[464] train's binary_logloss: 0.63593 valid's binary_logloss: 0.650751
bagging, val_score: 0.649063: 100%|##########| 10/10 [00:09<00:00, 1.01s/it][I 2020-09-27 04:47:27,517] Trial 36 finished with value: 0.6507510423009876 and parameters: {'bagging_fraction': 0.6850010940586274, 'bagging_freq': 3}. Best is trial 35 with value: 0.649063213862809.
bagging, val_score: 0.649063: 100%|##########| 10/10 [00:09<00:00, 1.02it/s]
feature_fraction_stage2, val_score: 0.649063: 0%| | 0/6 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005016 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656928 valid's binary_logloss: 0.661654
[200] train's binary_logloss: 0.646722 valid's binary_logloss: 0.652579
[300] train's binary_logloss: 0.641759 valid's binary_logloss: 0.65045
[400] train's binary_logloss: 0.638172 valid's binary_logloss: 0.649763
[500] train's binary_logloss: 0.634789 valid's binary_logloss: 0.649596
Early stopping, best iteration is:
[472] train's binary_logloss: 0.635712 valid's binary_logloss: 0.649063
feature_fraction_stage2, val_score: 0.649063: 17%|#6 | 1/6 [00:00<00:04, 1.11it/s][I 2020-09-27 04:47:28,434] Trial 37 finished with value: 0.649063213862809 and parameters: {'feature_fraction': 0.616}. Best is trial 37 with value: 0.649063213862809.
feature_fraction_stage2, val_score: 0.649063: 17%|#6 | 1/6 [00:00<00:04, 1.11it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004551 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656692 valid's binary_logloss: 0.661077
[200] train's binary_logloss: 0.646429 valid's binary_logloss: 0.652636
[300] train's binary_logloss: 0.641474 valid's binary_logloss: 0.650303
[400] train's binary_logloss: 0.637696 valid's binary_logloss: 0.649448
[500] train's binary_logloss: 0.634319 valid's binary_logloss: 0.64952
Early stopping, best iteration is:
[468] train's binary_logloss: 0.635429 valid's binary_logloss: 0.648813
feature_fraction_stage2, val_score: 0.648813: 33%|###3 | 2/6 [00:01<00:03, 1.11it/s][I 2020-09-27 04:47:29,319] Trial 38 finished with value: 0.6488132191521305 and parameters: {'feature_fraction': 0.6479999999999999}. Best is trial 38 with value: 0.6488132191521305.
feature_fraction_stage2, val_score: 0.648813: 33%|###3 | 2/6 [00:01<00:03, 1.11it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000576 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657062 valid's binary_logloss: 0.661671
[200] train's binary_logloss: 0.647032 valid's binary_logloss: 0.653351
[300] train's binary_logloss: 0.641928 valid's binary_logloss: 0.650494
[400] train's binary_logloss: 0.638384 valid's binary_logloss: 0.649678
[500] train's binary_logloss: 0.634981 valid's binary_logloss: 0.649656
[600] train's binary_logloss: 0.631952 valid's binary_logloss: 0.649533
Early stopping, best iteration is:
[551] train's binary_logloss: 0.633314 valid's binary_logloss: 0.649149
feature_fraction_stage2, val_score: 0.648813: 50%|##### | 3/6 [00:03<00:03, 1.06s/it][I 2020-09-27 04:47:30,759] Trial 39 finished with value: 0.6491494764152731 and parameters: {'feature_fraction': 0.552}. Best is trial 38 with value: 0.6488132191521305.
feature_fraction_stage2, val_score: 0.648813: 50%|##### | 3/6 [00:03<00:03, 1.06s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000576 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656892 valid's binary_logloss: 0.661978
[200] train's binary_logloss: 0.646753 valid's binary_logloss: 0.653932
[300] train's binary_logloss: 0.641708 valid's binary_logloss: 0.652504
[400] train's binary_logloss: 0.638016 valid's binary_logloss: 0.651
[500] train's binary_logloss: 0.634749 valid's binary_logloss: 0.651944
Early stopping, best iteration is:
[412] train's binary_logloss: 0.637606 valid's binary_logloss: 0.65077
feature_fraction_stage2, val_score: 0.648813: 67%|######6 | 4/6 [00:04<00:02, 1.00s/it][I 2020-09-27 04:47:31,621] Trial 40 finished with value: 0.6507703781531419 and parameters: {'feature_fraction': 0.584}. Best is trial 38 with value: 0.6488132191521305.
feature_fraction_stage2, val_score: 0.648813: 67%|######6 | 4/6 [00:04<00:02, 1.00s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004974 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656759 valid's binary_logloss: 0.661816
[200] train's binary_logloss: 0.646462 valid's binary_logloss: 0.653161
[300] train's binary_logloss: 0.641483 valid's binary_logloss: 0.650694
[400] train's binary_logloss: 0.6377 valid's binary_logloss: 0.649329
[500] train's binary_logloss: 0.634322 valid's binary_logloss: 0.649442
Early stopping, best iteration is:
[458] train's binary_logloss: 0.635675 valid's binary_logloss: 0.648669
feature_fraction_stage2, val_score: 0.648669: 83%|########3 | 5/6 [00:05<00:00, 1.01it/s][I 2020-09-27 04:47:32,586] Trial 41 finished with value: 0.6486693704765705 and parameters: {'feature_fraction': 0.6799999999999999}. Best is trial 41 with value: 0.6486693704765705.
feature_fraction_stage2, val_score: 0.648669: 83%|########3 | 5/6 [00:05<00:00, 1.01it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004631 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657062 valid's binary_logloss: 0.661671
[200] train's binary_logloss: 0.647032 valid's binary_logloss: 0.653351
[300] train's binary_logloss: 0.641928 valid's binary_logloss: 0.650494
[400] train's binary_logloss: 0.638384 valid's binary_logloss: 0.649678
[500] train's binary_logloss: 0.634981 valid's binary_logloss: 0.649656
[600] train's binary_logloss: 0.631952 valid's binary_logloss: 0.649533
Early stopping, best iteration is:
[551] train's binary_logloss: 0.633314 valid's binary_logloss: 0.649149
feature_fraction_stage2, val_score: 0.648669: 100%|##########| 6/6 [00:06<00:00, 1.11s/it][I 2020-09-27 04:47:33,975] Trial 42 finished with value: 0.6491494764152731 and parameters: {'feature_fraction': 0.52}. Best is trial 41 with value: 0.6486693704765705.
feature_fraction_stage2, val_score: 0.648669: 100%|##########| 6/6 [00:06<00:00, 1.08s/it]
regularization_factors, val_score: 0.648669: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012670 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656679 valid's binary_logloss: 0.661907
[200] train's binary_logloss: 0.646312 valid's binary_logloss: 0.653757
[300] train's binary_logloss: 0.641354 valid's binary_logloss: 0.651207
[400] train's binary_logloss: 0.637772 valid's binary_logloss: 0.650071
[500] train's binary_logloss: 0.634407 valid's binary_logloss: 0.649928
Early stopping, best iteration is:
[473] train's binary_logloss: 0.635291 valid's binary_logloss: 0.649607
regularization_factors, val_score: 0.648669: 5%|5 | 1/20 [00:01<00:20, 1.08s/it][I 2020-09-27 04:47:35,081] Trial 43 finished with value: 0.6496074604263665 and parameters: {'lambda_l1': 4.193685786846261e-08, 'lambda_l2': 0.3229663189729637}. Best is trial 43 with value: 0.6496074604263665.
regularization_factors, val_score: 0.648669: 5%|5 | 1/20 [00:01<00:20, 1.08s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011232 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657809 valid's binary_logloss: 0.663236
[200] train's binary_logloss: 0.647792 valid's binary_logloss: 0.654479
[300] train's binary_logloss: 0.643617 valid's binary_logloss: 0.650947
[400] train's binary_logloss: 0.640751 valid's binary_logloss: 0.650509
[500] train's binary_logloss: 0.638188 valid's binary_logloss: 0.650185
Early stopping, best iteration is:
[464] train's binary_logloss: 0.639104 valid's binary_logloss: 0.649722
regularization_factors, val_score: 0.648669: 10%|# | 2/20 [00:01<00:18, 1.03s/it][I 2020-09-27 04:47:35,996] Trial 44 finished with value: 0.6497218154257693 and parameters: {'lambda_l1': 7.661272890098566, 'lambda_l2': 1.6647681314940023e-07}. Best is trial 43 with value: 0.6496074604263665.
regularization_factors, val_score: 0.648669: 10%|# | 2/20 [00:02<00:18, 1.03s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005094 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656808 valid's binary_logloss: 0.661435
[200] train's binary_logloss: 0.646494 valid's binary_logloss: 0.653055
[300] train's binary_logloss: 0.641675 valid's binary_logloss: 0.650859
[400] train's binary_logloss: 0.637945 valid's binary_logloss: 0.649955
[500] train's binary_logloss: 0.634406 valid's binary_logloss: 0.649662
Early stopping, best iteration is:
[469] train's binary_logloss: 0.635491 valid's binary_logloss: 0.649128
regularization_factors, val_score: 0.648669: 15%|#5 | 3/20 [00:02<00:17, 1.01s/it][I 2020-09-27 04:47:36,939] Trial 45 finished with value: 0.6491279609766097 and parameters: {'lambda_l1': 0.038151596318754866, 'lambda_l2': 1.3493847157523479e-08}. Best is trial 45 with value: 0.6491279609766097.
regularization_factors, val_score: 0.648669: 15%|#5 | 3/20 [00:02<00:17, 1.01s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000809 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656763 valid's binary_logloss: 0.661526
[200] train's binary_logloss: 0.646359 valid's binary_logloss: 0.653842
[300] train's binary_logloss: 0.641324 valid's binary_logloss: 0.650914
[400] train's binary_logloss: 0.637727 valid's binary_logloss: 0.650544
[500] train's binary_logloss: 0.634443 valid's binary_logloss: 0.65063
Early stopping, best iteration is:
[457] train's binary_logloss: 0.635863 valid's binary_logloss: 0.649897
regularization_factors, val_score: 0.648669: 20%|## | 4/20 [00:04<00:17, 1.12s/it][I 2020-09-27 04:47:38,329] Trial 46 finished with value: 0.6498965604608956 and parameters: {'lambda_l1': 0.061724342064673816, 'lambda_l2': 1.1191098300700474e-08}. Best is trial 45 with value: 0.6491279609766097.
regularization_factors, val_score: 0.648669: 20%|## | 4/20 [00:04<00:17, 1.12s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005050 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656759 valid's binary_logloss: 0.661816
[200] train's binary_logloss: 0.646462 valid's binary_logloss: 0.653161
[300] train's binary_logloss: 0.641483 valid's binary_logloss: 0.650694
[400] train's binary_logloss: 0.637704 valid's binary_logloss: 0.649345
[500] train's binary_logloss: 0.634326 valid's binary_logloss: 0.649458
Early stopping, best iteration is:
[458] train's binary_logloss: 0.635679 valid's binary_logloss: 0.648685
regularization_factors, val_score: 0.648669: 25%|##5 | 5/20 [00:05<00:16, 1.09s/it][I 2020-09-27 04:47:39,333] Trial 47 finished with value: 0.6486854676681457 and parameters: {'lambda_l1': 7.797235764236721e-05, 'lambda_l2': 4.7617608412875624e-05}. Best is trial 47 with value: 0.6486854676681457.
regularization_factors, val_score: 0.648669: 25%|##5 | 5/20 [00:05<00:16, 1.09s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004751 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656759 valid's binary_logloss: 0.661816
[200] train's binary_logloss: 0.646462 valid's binary_logloss: 0.653161
[300] train's binary_logloss: 0.641483 valid's binary_logloss: 0.650694
[400] train's binary_logloss: 0.6377 valid's binary_logloss: 0.649329
[500] train's binary_logloss: 0.634322 valid's binary_logloss: 0.649442
Early stopping, best iteration is:
[458] train's binary_logloss: 0.635675 valid's binary_logloss: 0.648669
regularization_factors, val_score: 0.648669: 30%|### | 6/20 [00:06<00:14, 1.03s/it][I 2020-09-27 04:47:40,228] Trial 48 finished with value: 0.6486693564289175 and parameters: {'lambda_l1': 4.6907860757180705e-07, 'lambda_l2': 0.0001968190366629125}. Best is trial 48 with value: 0.6486693564289175.
regularization_factors, val_score: 0.648669: 30%|### | 6/20 [00:06<00:14, 1.03s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000809 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656759 valid's binary_logloss: 0.661816
[200] train's binary_logloss: 0.646461 valid's binary_logloss: 0.653161
[300] train's binary_logloss: 0.641483 valid's binary_logloss: 0.650694
[400] train's binary_logloss: 0.637703 valid's binary_logloss: 0.649345
[500] train's binary_logloss: 0.634325 valid's binary_logloss: 0.649458
Early stopping, best iteration is:
[458] train's binary_logloss: 0.635679 valid's binary_logloss: 0.648685
regularization_factors, val_score: 0.648669: 35%|###5 | 7/20 [00:07<00:12, 1.01it/s][I 2020-09-27 04:47:41,143] Trial 49 finished with value: 0.6486853650384133 and parameters: {'lambda_l1': 7.106958303444078e-07, 'lambda_l2': 0.00031503172943699124}. Best is trial 48 with value: 0.6486693564289175.
regularization_factors, val_score: 0.648669: 35%|###5 | 7/20 [00:07<00:12, 1.01it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004879 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656759 valid's binary_logloss: 0.661816
[200] train's binary_logloss: 0.646462 valid's binary_logloss: 0.653161
[300] train's binary_logloss: 0.641483 valid's binary_logloss: 0.650694
[400] train's binary_logloss: 0.637704 valid's binary_logloss: 0.649345
[500] train's binary_logloss: 0.634326 valid's binary_logloss: 0.649458
Early stopping, best iteration is:
[458] train's binary_logloss: 0.635679 valid's binary_logloss: 0.648685
regularization_factors, val_score: 0.648669: 40%|#### | 8/20 [00:08<00:13, 1.09s/it][I 2020-09-27 04:47:42,447] Trial 50 finished with value: 0.6486854411094597 and parameters: {'lambda_l1': 4.4792617037779606e-07, 'lambda_l2': 0.00040119958572656174}. Best is trial 48 with value: 0.6486693564289175.
regularization_factors, val_score: 0.648669: 40%|#### | 8/20 [00:08<00:13, 1.09s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000826 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656759 valid's binary_logloss: 0.661816
[200] train's binary_logloss: 0.646462 valid's binary_logloss: 0.653161
[300] train's binary_logloss: 0.641483 valid's binary_logloss: 0.650694
[400] train's binary_logloss: 0.6377 valid's binary_logloss: 0.649329
[500] train's binary_logloss: 0.634322 valid's binary_logloss: 0.649442
Early stopping, best iteration is:
[458] train's binary_logloss: 0.635675 valid's binary_logloss: 0.648669
regularization_factors, val_score: 0.648669: 45%|####5 | 9/20 [00:09<00:11, 1.05s/it][I 2020-09-27 04:47:43,405] Trial 51 finished with value: 0.6486693484397528 and parameters: {'lambda_l1': 3.4434140107307795e-07, 'lambda_l2': 0.00030839964549718207}. Best is trial 51 with value: 0.6486693484397528.
regularization_factors, val_score: 0.648669: 45%|####5 | 9/20 [00:09<00:11, 1.05s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000859 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656759 valid's binary_logloss: 0.661816
[200] train's binary_logloss: 0.646462 valid's binary_logloss: 0.653161
[300] train's binary_logloss: 0.641483 valid's binary_logloss: 0.650694
[400] train's binary_logloss: 0.637704 valid's binary_logloss: 0.649345
[500] train's binary_logloss: 0.634326 valid's binary_logloss: 0.649458
Early stopping, best iteration is:
[458] train's binary_logloss: 0.635679 valid's binary_logloss: 0.648685
regularization_factors, val_score: 0.648669: 50%|##### | 10/20 [00:10<00:10, 1.01s/it][I 2020-09-27 04:47:44,335] Trial 52 finished with value: 0.6486854548458479 and parameters: {'lambda_l1': 2.9104522149216847e-07, 'lambda_l2': 0.00021910741939920448}. Best is trial 51 with value: 0.6486693484397528.
regularization_factors, val_score: 0.648669: 50%|##### | 10/20 [00:10<00:10, 1.01s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009619 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656759 valid's binary_logloss: 0.661816
[200] train's binary_logloss: 0.646462 valid's binary_logloss: 0.653161
[300] train's binary_logloss: 0.641483 valid's binary_logloss: 0.650694
[400] train's binary_logloss: 0.637704 valid's binary_logloss: 0.649345
[500] train's binary_logloss: 0.634326 valid's binary_logloss: 0.649458
Early stopping, best iteration is:
[458] train's binary_logloss: 0.635679 valid's binary_logloss: 0.648685
regularization_factors, val_score: 0.648669: 55%|#####5 | 11/20 [00:11<00:09, 1.03s/it][I 2020-09-27 04:47:45,395] Trial 53 finished with value: 0.6486854423114767 and parameters: {'lambda_l1': 1.8051040240033813e-07, 'lambda_l2': 0.0003854090838672508}. Best is trial 51 with value: 0.6486693484397528.
regularization_factors, val_score: 0.648669: 55%|#####5 | 11/20 [00:11<00:09, 1.03s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002587 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656759 valid's binary_logloss: 0.661816
[200] train's binary_logloss: 0.646462 valid's binary_logloss: 0.653161
[300] train's binary_logloss: 0.641483 valid's binary_logloss: 0.650694
[400] train's binary_logloss: 0.637704 valid's binary_logloss: 0.649345
[500] train's binary_logloss: 0.634326 valid's binary_logloss: 0.649458
Early stopping, best iteration is:
[458] train's binary_logloss: 0.635679 valid's binary_logloss: 0.648685
regularization_factors, val_score: 0.648669: 60%|###### | 12/20 [00:12<00:08, 1.11s/it][I 2020-09-27 04:47:46,692] Trial 54 finished with value: 0.6486853768130912 and parameters: {'lambda_l1': 1.238928167433891e-06, 'lambda_l2': 0.001256677247104136}. Best is trial 51 with value: 0.6486693484397528.
regularization_factors, val_score: 0.648669: 60%|###### | 12/20 [00:12<00:08, 1.11s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009674 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656759 valid's binary_logloss: 0.661816
[200] train's binary_logloss: 0.646462 valid's binary_logloss: 0.653161
[300] train's binary_logloss: 0.641484 valid's binary_logloss: 0.650694
[400] train's binary_logloss: 0.637701 valid's binary_logloss: 0.649329
[500] train's binary_logloss: 0.634323 valid's binary_logloss: 0.649442
Early stopping, best iteration is:
[458] train's binary_logloss: 0.635676 valid's binary_logloss: 0.648669
regularization_factors, val_score: 0.648669: 65%|######5 | 13/20 [00:13<00:07, 1.05s/it][I 2020-09-27 04:47:47,595] Trial 55 finished with value: 0.6486691459174234 and parameters: {'lambda_l1': 4.463187899572695e-06, 'lambda_l2': 0.0031440113191708815}. Best is trial 55 with value: 0.6486691459174234.
regularization_factors, val_score: 0.648669: 65%|######5 | 13/20 [00:13<00:07, 1.05s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004983 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656759 valid's binary_logloss: 0.661816
[200] train's binary_logloss: 0.646463 valid's binary_logloss: 0.653161
[300] train's binary_logloss: 0.641496 valid's binary_logloss: 0.650455
[400] train's binary_logloss: 0.637726 valid's binary_logloss: 0.649177
[500] train's binary_logloss: 0.634323 valid's binary_logloss: 0.649235
Early stopping, best iteration is:
[458] train's binary_logloss: 0.635717 valid's binary_logloss: 0.648475
regularization_factors, val_score: 0.648475: 70%|####### | 14/20 [00:14<00:06, 1.00s/it][I 2020-09-27 04:47:48,497] Trial 56 finished with value: 0.6484745400508054 and parameters: {'lambda_l1': 9.72103845454671e-06, 'lambda_l2': 0.011377674391625028}. Best is trial 56 with value: 0.6484745400508054.
regularization_factors, val_score: 0.648475: 70%|####### | 14/20 [00:14<00:06, 1.00s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001108 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656804 valid's binary_logloss: 0.661432
[200] train's binary_logloss: 0.646512 valid's binary_logloss: 0.653055
[300] train's binary_logloss: 0.641411 valid's binary_logloss: 0.650947
[400] train's binary_logloss: 0.637723 valid's binary_logloss: 0.649987
[500] train's binary_logloss: 0.634326 valid's binary_logloss: 0.650052
Early stopping, best iteration is:
[464] train's binary_logloss: 0.635548 valid's binary_logloss: 0.649488
regularization_factors, val_score: 0.648475: 75%|#######5 | 15/20 [00:15<00:05, 1.08s/it][I 2020-09-27 04:47:49,768] Trial 57 finished with value: 0.6494877361955586 and parameters: {'lambda_l1': 1.3431278103969494e-05, 'lambda_l2': 0.02447854418046466}. Best is trial 56 with value: 0.6484745400508054.
regularization_factors, val_score: 0.648475: 75%|#######5 | 15/20 [00:15<00:05, 1.08s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.015062 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656759 valid's binary_logloss: 0.661816
[200] train's binary_logloss: 0.646462 valid's binary_logloss: 0.653161
[300] train's binary_logloss: 0.641495 valid's binary_logloss: 0.650455
[400] train's binary_logloss: 0.637725 valid's binary_logloss: 0.649177
[500] train's binary_logloss: 0.634322 valid's binary_logloss: 0.649235
Early stopping, best iteration is:
[458] train's binary_logloss: 0.635715 valid's binary_logloss: 0.648475
regularization_factors, val_score: 0.648475: 80%|######## | 16/20 [00:16<00:04, 1.08s/it][I 2020-09-27 04:47:50,844] Trial 58 finished with value: 0.6484747613620405 and parameters: {'lambda_l1': 5.558047158433351e-06, 'lambda_l2': 0.007068676788026479}. Best is trial 56 with value: 0.6484745400508054.
regularization_factors, val_score: 0.648475: 80%|######## | 16/20 [00:16<00:04, 1.08s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000877 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656759 valid's binary_logloss: 0.661817
[200] train's binary_logloss: 0.646463 valid's binary_logloss: 0.653162
[300] train's binary_logloss: 0.641497 valid's binary_logloss: 0.650455
[400] train's binary_logloss: 0.637728 valid's binary_logloss: 0.649176
[500] train's binary_logloss: 0.634326 valid's binary_logloss: 0.649234
Early stopping, best iteration is:
[458] train's binary_logloss: 0.635719 valid's binary_logloss: 0.648474
regularization_factors, val_score: 0.648474: 85%|########5 | 17/20 [00:17<00:03, 1.09s/it][I 2020-09-27 04:47:51,940] Trial 59 finished with value: 0.6484741879925516 and parameters: {'lambda_l1': 1.1402866509712536e-05, 'lambda_l2': 0.018245531549826995}. Best is trial 59 with value: 0.6484741879925516.
regularization_factors, val_score: 0.648474: 85%|########5 | 17/20 [00:17<00:03, 1.09s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004790 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656759 valid's binary_logloss: 0.661817
[200] train's binary_logloss: 0.646463 valid's binary_logloss: 0.653162
[300] train's binary_logloss: 0.641497 valid's binary_logloss: 0.650455
[400] train's binary_logloss: 0.637728 valid's binary_logloss: 0.649176
[500] train's binary_logloss: 0.634326 valid's binary_logloss: 0.649234
Early stopping, best iteration is:
[458] train's binary_logloss: 0.635719 valid's binary_logloss: 0.648474
regularization_factors, val_score: 0.648474: 90%|######### | 18/20 [00:18<00:02, 1.03s/it][I 2020-09-27 04:47:52,825] Trial 60 finished with value: 0.6484741937796707 and parameters: {'lambda_l1': 1.9414476880011605e-05, 'lambda_l2': 0.018137780474702055}. Best is trial 59 with value: 0.6484741879925516.
regularization_factors, val_score: 0.648474: 90%|######### | 18/20 [00:18<00:02, 1.03s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004872 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656759 valid's binary_logloss: 0.661816
[200] train's binary_logloss: 0.646463 valid's binary_logloss: 0.653161
[300] train's binary_logloss: 0.641496 valid's binary_logloss: 0.650455
[400] train's binary_logloss: 0.637727 valid's binary_logloss: 0.649177
[500] train's binary_logloss: 0.634324 valid's binary_logloss: 0.649235
Early stopping, best iteration is:
[458] train's binary_logloss: 0.635717 valid's binary_logloss: 0.648474
regularization_factors, val_score: 0.648474: 95%|#########5| 19/20 [00:20<00:01, 1.14s/it][I 2020-09-27 04:47:54,216] Trial 61 finished with value: 0.6484744650230337 and parameters: {'lambda_l1': 1.6059055968360456e-05, 'lambda_l2': 0.012844094468385317}. Best is trial 59 with value: 0.6484741879925516.
regularization_factors, val_score: 0.648474: 95%|#########5| 19/20 [00:20<00:01, 1.14s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004738 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.65676 valid's binary_logloss: 0.661817
[200] train's binary_logloss: 0.646463 valid's binary_logloss: 0.653162
[300] train's binary_logloss: 0.641498 valid's binary_logloss: 0.650454
[400] train's binary_logloss: 0.637729 valid's binary_logloss: 0.649176
[500] train's binary_logloss: 0.634329 valid's binary_logloss: 0.649256
Early stopping, best iteration is:
[458] train's binary_logloss: 0.63572 valid's binary_logloss: 0.648474
regularization_factors, val_score: 0.648474: 100%|##########| 20/20 [00:21<00:00, 1.09s/it][I 2020-09-27 04:47:55,185] Trial 62 finished with value: 0.6484739197241262 and parameters: {'lambda_l1': 1.4740300654925883e-05, 'lambda_l2': 0.02349270671135121}. Best is trial 62 with value: 0.6484739197241262.
regularization_factors, val_score: 0.648474: 100%|##########| 20/20 [00:21<00:00, 1.06s/it]
min_data_in_leaf, val_score: 0.648474: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000891 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.65676 valid's binary_logloss: 0.661817
[200] train's binary_logloss: 0.646525 valid's binary_logloss: 0.653595
[300] train's binary_logloss: 0.641846 valid's binary_logloss: 0.650251
[400] train's binary_logloss: 0.638553 valid's binary_logloss: 0.649577
[500] train's binary_logloss: 0.63552 valid's binary_logloss: 0.64935
Early stopping, best iteration is:
[464] train's binary_logloss: 0.636644 valid's binary_logloss: 0.648879
min_data_in_leaf, val_score: 0.648474: 20%|## | 1/5 [00:01<00:04, 1.01s/it][I 2020-09-27 04:47:56,208] Trial 63 finished with value: 0.648879477038123 and parameters: {'min_child_samples': 100}. Best is trial 63 with value: 0.648879477038123.
min_data_in_leaf, val_score: 0.648474: 20%|## | 1/5 [00:01<00:04, 1.01s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010338 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.65676 valid's binary_logloss: 0.661817
[200] train's binary_logloss: 0.646463 valid's binary_logloss: 0.653162
[300] train's binary_logloss: 0.6415 valid's binary_logloss: 0.650452
[400] train's binary_logloss: 0.637791 valid's binary_logloss: 0.649375
[500] train's binary_logloss: 0.634508 valid's binary_logloss: 0.649431
Early stopping, best iteration is:
[451] train's binary_logloss: 0.636075 valid's binary_logloss: 0.648808
min_data_in_leaf, val_score: 0.648474: 40%|#### | 2/5 [00:02<00:03, 1.03s/it][I 2020-09-27 04:47:57,286] Trial 64 finished with value: 0.6488082911033146 and parameters: {'min_child_samples': 25}. Best is trial 64 with value: 0.6488082911033146.
min_data_in_leaf, val_score: 0.648474: 40%|#### | 2/5 [00:02<00:03, 1.03s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.015161 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.65676 valid's binary_logloss: 0.661817
[200] train's binary_logloss: 0.646493 valid's binary_logloss: 0.65325
[300] train's binary_logloss: 0.641497 valid's binary_logloss: 0.650822
[400] train's binary_logloss: 0.637788 valid's binary_logloss: 0.649913
[500] train's binary_logloss: 0.634246 valid's binary_logloss: 0.649666
Early stopping, best iteration is:
[468] train's binary_logloss: 0.635328 valid's binary_logloss: 0.649123
min_data_in_leaf, val_score: 0.648474: 60%|###### | 3/5 [00:03<00:02, 1.10s/it][I 2020-09-27 04:47:58,535] Trial 65 finished with value: 0.649122937898972 and parameters: {'min_child_samples': 10}. Best is trial 64 with value: 0.6488082911033146.
min_data_in_leaf, val_score: 0.648474: 60%|###### | 3/5 [00:03<00:02, 1.10s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001909 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.65676 valid's binary_logloss: 0.661817
[200] train's binary_logloss: 0.64648 valid's binary_logloss: 0.653437
[300] train's binary_logloss: 0.641645 valid's binary_logloss: 0.650437
[400] train's binary_logloss: 0.638049 valid's binary_logloss: 0.649312
[500] train's binary_logloss: 0.634894 valid's binary_logloss: 0.649312
Early stopping, best iteration is:
[447] train's binary_logloss: 0.63654 valid's binary_logloss: 0.648959
min_data_in_leaf, val_score: 0.648474: 80%|######## | 4/5 [00:04<00:01, 1.04s/it][I 2020-09-27 04:47:59,458] Trial 66 finished with value: 0.6489591737909831 and parameters: {'min_child_samples': 50}. Best is trial 64 with value: 0.6488082911033146.
min_data_in_leaf, val_score: 0.648474: 80%|######## | 4/5 [00:04<00:01, 1.04s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001493 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4242
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309
[LightGBM] [Info] Start training from score -0.022309
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.65676 valid's binary_logloss: 0.661817
[200] train's binary_logloss: 0.646493 valid's binary_logloss: 0.65325
[300] train's binary_logloss: 0.64151 valid's binary_logloss: 0.65102
[400] train's binary_logloss: 0.637718 valid's binary_logloss: 0.650205
[500] train's binary_logloss: 0.634224 valid's binary_logloss: 0.649812
Early stopping, best iteration is:
[459] train's binary_logloss: 0.635631 valid's binary_logloss: 0.649459
min_data_in_leaf, val_score: 0.648474: 100%|##########| 5/5 [00:05<00:00, 1.02s/it][I 2020-09-27 04:48:00,406] Trial 67 finished with value: 0.6494585505180116 and parameters: {'min_child_samples': 5}. Best is trial 64 with value: 0.6488082911033146.
min_data_in_leaf, val_score: 0.648474: 100%|##########| 5/5 [00:05<00:00, 1.04s/it]
Fold : 9
[I 2020-09-27 04:48:00,469] A new study created in memory with name: no-name-7b517b46-d774-4255-8991-4b95254b4cf4
feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000996 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.571415 valid's binary_logloss: 0.664205
Early stopping, best iteration is:
[56] train's binary_logloss: 0.602615 valid's binary_logloss: 0.661775
feature_fraction, val_score: 0.661775: 14%|#4 | 1/7 [00:00<00:05, 1.12it/s][I 2020-09-27 04:48:01,375] Trial 0 finished with value: 0.6617754129584598 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 0 with value: 0.6617754129584598.
feature_fraction, val_score: 0.661775: 14%|#4 | 1/7 [00:00<00:05, 1.12it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001704 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.577907 valid's binary_logloss: 0.66332
Early stopping, best iteration is:
[71] train's binary_logloss: 0.596884 valid's binary_logloss: 0.662116
feature_fraction, val_score: 0.661775: 29%|##8 | 2/7 [00:02<00:04, 1.04it/s][I 2020-09-27 04:48:02,518] Trial 1 finished with value: 0.6621160574043524 and parameters: {'feature_fraction': 0.5}. Best is trial 0 with value: 0.6617754129584598.
feature_fraction, val_score: 0.661775: 29%|##8 | 2/7 [00:02<00:04, 1.04it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000446 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.581403 valid's binary_logloss: 0.664323
Early stopping, best iteration is:
[69] train's binary_logloss: 0.601689 valid's binary_logloss: 0.663133
feature_fraction, val_score: 0.661775: 43%|####2 | 3/7 [00:02<00:03, 1.21it/s][I 2020-09-27 04:48:03,026] Trial 2 finished with value: 0.6631333345540273 and parameters: {'feature_fraction': 0.4}. Best is trial 0 with value: 0.6617754129584598.
feature_fraction, val_score: 0.661775: 43%|####2 | 3/7 [00:02<00:03, 1.21it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000926 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.569848 valid's binary_logloss: 0.664707
Early stopping, best iteration is:
[67] train's binary_logloss: 0.59321 valid's binary_logloss: 0.662756
feature_fraction, val_score: 0.661775: 57%|#####7 | 4/7 [00:03<00:02, 1.35it/s][I 2020-09-27 04:48:03,555] Trial 3 finished with value: 0.6627562719479632 and parameters: {'feature_fraction': 1.0}. Best is trial 0 with value: 0.6617754129584598.
feature_fraction, val_score: 0.661775: 57%|#####7 | 4/7 [00:03<00:02, 1.35it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000814 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.573306 valid's binary_logloss: 0.661877
[200] train's binary_logloss: 0.517495 valid's binary_logloss: 0.664844
Early stopping, best iteration is:
[113] train's binary_logloss: 0.565412 valid's binary_logloss: 0.660855
feature_fraction, val_score: 0.660855: 71%|#######1 | 5/7 [00:03<00:01, 1.42it/s][I 2020-09-27 04:48:04,180] Trial 4 finished with value: 0.6608549078376256 and parameters: {'feature_fraction': 0.8}. Best is trial 4 with value: 0.6608549078376256.
feature_fraction, val_score: 0.660855: 71%|#######1 | 5/7 [00:03<00:01, 1.42it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004908 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.574579 valid's binary_logloss: 0.662991
Early stopping, best iteration is:
[54] train's binary_logloss: 0.606825 valid's binary_logloss: 0.661109
feature_fraction, val_score: 0.660855: 86%|########5 | 6/7 [00:04<00:00, 1.58it/s][I 2020-09-27 04:48:04,649] Trial 5 finished with value: 0.66110942300072 and parameters: {'feature_fraction': 0.7}. Best is trial 4 with value: 0.6608549078376256.
feature_fraction, val_score: 0.660855: 86%|########5 | 6/7 [00:04<00:00, 1.58it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004781 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.575121 valid's binary_logloss: 0.665199
Early stopping, best iteration is:
[58] train's binary_logloss: 0.604585 valid's binary_logloss: 0.663434
feature_fraction, val_score: 0.660855: 100%|##########| 7/7 [00:04<00:00, 1.58it/s][I 2020-09-27 04:48:05,278] Trial 6 finished with value: 0.6634341152776219 and parameters: {'feature_fraction': 0.6}. Best is trial 4 with value: 0.6608549078376256.
feature_fraction, val_score: 0.660855: 100%|##########| 7/7 [00:04<00:00, 1.46it/s]
num_leaves, val_score: 0.660855: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004289 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.573306 valid's binary_logloss: 0.661877
[200] train's binary_logloss: 0.517495 valid's binary_logloss: 0.664844
Early stopping, best iteration is:
[113] train's binary_logloss: 0.565412 valid's binary_logloss: 0.660855
num_leaves, val_score: 0.660855: 5%|5 | 1/20 [00:01<00:37, 1.99s/it][I 2020-09-27 04:48:07,289] Trial 7 finished with value: 0.6608549078376256 and parameters: {'num_leaves': 31}. Best is trial 7 with value: 0.6608549078376256.
num_leaves, val_score: 0.660855: 5%|5 | 1/20 [00:02<00:37, 1.99s/it][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.009560 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.291282 valid's binary_logloss: 0.683399
Early stopping, best iteration is:
[25] train's binary_logloss: 0.515751 valid's binary_logloss: 0.662419
num_leaves, val_score: 0.660855: 10%|# | 2/20 [00:03<00:32, 1.79s/it][I 2020-09-27 04:48:08,615] Trial 8 finished with value: 0.6624189569421138 and parameters: {'num_leaves': 227}. Best is trial 7 with value: 0.6608549078376256.
num_leaves, val_score: 0.660855: 10%|# | 2/20 [00:03<00:32, 1.79s/it][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000898 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.633793 valid's binary_logloss: 0.66169
[200] train's binary_logloss: 0.616176 valid's binary_logloss: 0.662116
Early stopping, best iteration is:
[127] train's binary_logloss: 0.628472 valid's binary_logloss: 0.661027
num_leaves, val_score: 0.660855: 15%|#5 | 3/20 [00:03<00:24, 1.41s/it][I 2020-09-27 04:48:09,146] Trial 9 finished with value: 0.6610265822259738 and parameters: {'num_leaves': 8}. Best is trial 7 with value: 0.6608549078376256.
num_leaves, val_score: 0.660855: 15%|#5 | 3/20 [00:03<00:24, 1.41s/it][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000937 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.277955 valid's binary_logloss: 0.68283
Early stopping, best iteration is:
[29] train's binary_logloss: 0.489512 valid's binary_logloss: 0.667966
num_leaves, val_score: 0.660855: 20%|## | 4/20 [00:05<00:24, 1.52s/it][I 2020-09-27 04:48:10,921] Trial 10 finished with value: 0.6679657755747807 and parameters: {'num_leaves': 241}. Best is trial 7 with value: 0.6608549078376256.
num_leaves, val_score: 0.660855: 20%|## | 4/20 [00:05<00:24, 1.52s/it][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000991 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.586674 valid's binary_logloss: 0.66365
Early stopping, best iteration is:
[74] train's binary_logloss: 0.601053 valid's binary_logloss: 0.662634
num_leaves, val_score: 0.660855: 25%|##5 | 5/20 [00:06<00:18, 1.21s/it][I 2020-09-27 04:48:11,417] Trial 11 finished with value: 0.6626335832631651 and parameters: {'num_leaves': 25}. Best is trial 7 with value: 0.6608549078376256.
num_leaves, val_score: 0.660855: 25%|##5 | 5/20 [00:06<00:18, 1.21s/it][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000938 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.437791 valid's binary_logloss: 0.672762
Early stopping, best iteration is:
[35] train's binary_logloss: 0.555775 valid's binary_logloss: 0.664193
num_leaves, val_score: 0.660855: 30%|### | 6/20 [00:06<00:15, 1.07s/it][I 2020-09-27 04:48:12,160] Trial 12 finished with value: 0.664192632015902 and parameters: {'num_leaves': 104}. Best is trial 7 with value: 0.6608549078376256.
num_leaves, val_score: 0.660855: 30%|### | 6/20 [00:06<00:15, 1.07s/it][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000963 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.458686 valid's binary_logloss: 0.668395
Early stopping, best iteration is:
[40] train's binary_logloss: 0.556167 valid's binary_logloss: 0.664721
num_leaves, val_score: 0.660855: 35%|###5 | 7/20 [00:07<00:12, 1.03it/s][I 2020-09-27 04:48:12,890] Trial 13 finished with value: 0.6647214922013416 and parameters: {'num_leaves': 90}. Best is trial 7 with value: 0.6608549078376256.
num_leaves, val_score: 0.660855: 35%|###5 | 7/20 [00:07<00:12, 1.03it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009252 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.350518 valid's binary_logloss: 0.684373
Early stopping, best iteration is:
[22] train's binary_logloss: 0.558627 valid's binary_logloss: 0.66659
num_leaves, val_score: 0.660855: 40%|#### | 8/20 [00:08<00:11, 1.07it/s][I 2020-09-27 04:48:13,742] Trial 14 finished with value: 0.6665899695534304 and parameters: {'num_leaves': 167}. Best is trial 7 with value: 0.6608549078376256.
num_leaves, val_score: 0.660855: 40%|#### | 8/20 [00:08<00:11, 1.07it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004731 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.532376 valid's binary_logloss: 0.669923
Early stopping, best iteration is:
[40] train's binary_logloss: 0.596466 valid's binary_logloss: 0.665493
num_leaves, val_score: 0.660855: 45%|####5 | 9/20 [00:09<00:10, 1.04it/s][I 2020-09-27 04:48:14,759] Trial 15 finished with value: 0.665493198325232 and parameters: {'num_leaves': 50}. Best is trial 7 with value: 0.6608549078376256.
num_leaves, val_score: 0.660855: 45%|####5 | 9/20 [00:09<00:10, 1.04it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000797 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.350518 valid's binary_logloss: 0.684373
Early stopping, best iteration is:
[22] train's binary_logloss: 0.558627 valid's binary_logloss: 0.66659
num_leaves, val_score: 0.660855: 50%|##### | 10/20 [00:10<00:09, 1.04it/s][I 2020-09-27 04:48:15,722] Trial 16 finished with value: 0.6665899695534304 and parameters: {'num_leaves': 167}. Best is trial 7 with value: 0.6608549078376256.
num_leaves, val_score: 0.660855: 50%|##### | 10/20 [00:10<00:09, 1.04it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004981 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.512177 valid's binary_logloss: 0.666645
Early stopping, best iteration is:
[29] train's binary_logloss: 0.60465 valid's binary_logloss: 0.662506
num_leaves, val_score: 0.660855: 55%|#####5 | 11/20 [00:10<00:07, 1.20it/s][I 2020-09-27 04:48:16,266] Trial 17 finished with value: 0.6625059680700698 and parameters: {'num_leaves': 60}. Best is trial 7 with value: 0.6608549078376256.
num_leaves, val_score: 0.660855: 55%|#####5 | 11/20 [00:10<00:07, 1.20it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001403 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650402 valid's binary_logloss: 0.665088
[200] train's binary_logloss: 0.639194 valid's binary_logloss: 0.660755
[300] train's binary_logloss: 0.632319 valid's binary_logloss: 0.66028
[400] train's binary_logloss: 0.626788 valid's binary_logloss: 0.660425
Early stopping, best iteration is:
[360] train's binary_logloss: 0.628825 valid's binary_logloss: 0.659947
num_leaves, val_score: 0.659947: 60%|###### | 12/20 [00:11<00:06, 1.22it/s][I 2020-09-27 04:48:17,046] Trial 18 finished with value: 0.659946568502427 and parameters: {'num_leaves': 4}. Best is trial 18 with value: 0.659946568502427.
num_leaves, val_score: 0.659947: 60%|###### | 12/20 [00:11<00:06, 1.22it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001058 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.637047 valid's binary_logloss: 0.663485
[200] train's binary_logloss: 0.621339 valid's binary_logloss: 0.662464
[300] train's binary_logloss: 0.608631 valid's binary_logloss: 0.662754
Early stopping, best iteration is:
[218] train's binary_logloss: 0.618989 valid's binary_logloss: 0.661824
num_leaves, val_score: 0.659947: 65%|######5 | 13/20 [00:12<00:05, 1.31it/s][I 2020-09-27 04:48:17,672] Trial 19 finished with value: 0.6618235788394593 and parameters: {'num_leaves': 7}. Best is trial 18 with value: 0.659946568502427.
num_leaves, val_score: 0.659947: 65%|######5 | 13/20 [00:12<00:05, 1.31it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.006735 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657096 valid's binary_logloss: 0.667367
[200] train's binary_logloss: 0.646565 valid's binary_logloss: 0.661837
[300] train's binary_logloss: 0.641199 valid's binary_logloss: 0.660556
[400] train's binary_logloss: 0.637372 valid's binary_logloss: 0.660306
Early stopping, best iteration is:
[357] train's binary_logloss: 0.638883 valid's binary_logloss: 0.659984
num_leaves, val_score: 0.659947: 70%|####### | 14/20 [00:13<00:05, 1.13it/s][I 2020-09-27 04:48:18,847] Trial 20 finished with value: 0.6599838058476127 and parameters: {'num_leaves': 3}. Best is trial 18 with value: 0.659946568502427.
num_leaves, val_score: 0.659947: 70%|####### | 14/20 [00:13<00:05, 1.13it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001022 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.645096 valid's binary_logloss: 0.662705
[200] train's binary_logloss: 0.632688 valid's binary_logloss: 0.660038
Early stopping, best iteration is:
[193] train's binary_logloss: 0.633406 valid's binary_logloss: 0.660001
num_leaves, val_score: 0.659947: 75%|#######5 | 15/20 [00:14<00:03, 1.26it/s][I 2020-09-27 04:48:19,418] Trial 21 finished with value: 0.6600011118675012 and parameters: {'num_leaves': 5}. Best is trial 18 with value: 0.659946568502427.
num_leaves, val_score: 0.659947: 75%|#######5 | 15/20 [00:14<00:03, 1.26it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001022 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.629941 valid's binary_logloss: 0.66087
[200] train's binary_logloss: 0.610687 valid's binary_logloss: 0.662744
Early stopping, best iteration is:
[125] train's binary_logloss: 0.624522 valid's binary_logloss: 0.66051
num_leaves, val_score: 0.659947: 80%|######## | 16/20 [00:14<00:02, 1.38it/s][I 2020-09-27 04:48:19,995] Trial 22 finished with value: 0.6605096965777126 and parameters: {'num_leaves': 9}. Best is trial 18 with value: 0.659946568502427.
num_leaves, val_score: 0.659947: 80%|######## | 16/20 [00:14<00:02, 1.38it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.006817 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.512177 valid's binary_logloss: 0.666645
Early stopping, best iteration is:
[29] train's binary_logloss: 0.60465 valid's binary_logloss: 0.662506
num_leaves, val_score: 0.659947: 85%|########5 | 17/20 [00:15<00:02, 1.48it/s][I 2020-09-27 04:48:20,555] Trial 23 finished with value: 0.6625059680700698 and parameters: {'num_leaves': 60}. Best is trial 18 with value: 0.659946568502427.
num_leaves, val_score: 0.659947: 85%|########5 | 17/20 [00:15<00:02, 1.48it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001075 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650402 valid's binary_logloss: 0.665088
[200] train's binary_logloss: 0.639194 valid's binary_logloss: 0.660755
[300] train's binary_logloss: 0.632319 valid's binary_logloss: 0.66028
[400] train's binary_logloss: 0.626788 valid's binary_logloss: 0.660425
Early stopping, best iteration is:
[360] train's binary_logloss: 0.628825 valid's binary_logloss: 0.659947
num_leaves, val_score: 0.659947: 90%|######### | 18/20 [00:16<00:01, 1.40it/s][I 2020-09-27 04:48:21,359] Trial 24 finished with value: 0.6599465685024268 and parameters: {'num_leaves': 4}. Best is trial 24 with value: 0.6599465685024268.
num_leaves, val_score: 0.659947: 90%|######### | 18/20 [00:16<00:01, 1.40it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000833 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.569877 valid's binary_logloss: 0.667063
Early stopping, best iteration is:
[43] train's binary_logloss: 0.613515 valid's binary_logloss: 0.663355
num_leaves, val_score: 0.659947: 95%|#########5| 19/20 [00:17<00:00, 1.27it/s][I 2020-09-27 04:48:22,313] Trial 25 finished with value: 0.6633553868179578 and parameters: {'num_leaves': 32}. Best is trial 24 with value: 0.6599465685024268.
num_leaves, val_score: 0.659947: 95%|#########5| 19/20 [00:17<00:00, 1.27it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000762 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.466555 valid's binary_logloss: 0.674499
Early stopping, best iteration is:
[32] train's binary_logloss: 0.577932 valid's binary_logloss: 0.665241
num_leaves, val_score: 0.659947: 100%|##########| 20/20 [00:17<00:00, 1.30it/s][I 2020-09-27 04:48:23,045] Trial 26 finished with value: 0.6652408810612898 and parameters: {'num_leaves': 85}. Best is trial 24 with value: 0.6599465685024268.
num_leaves, val_score: 0.659947: 100%|##########| 20/20 [00:17<00:00, 1.13it/s]
bagging, val_score: 0.659947: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000831 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.649887 valid's binary_logloss: 0.663942
[200] train's binary_logloss: 0.639122 valid's binary_logloss: 0.660082
[300] train's binary_logloss: 0.632108 valid's binary_logloss: 0.660084
Early stopping, best iteration is:
[239] train's binary_logloss: 0.636191 valid's binary_logloss: 0.659231
bagging, val_score: 0.659231: 10%|# | 1/10 [00:00<00:06, 1.48it/s][I 2020-09-27 04:48:23,737] Trial 27 finished with value: 0.6592308295470516 and parameters: {'bagging_fraction': 0.8273856061488931, 'bagging_freq': 3}. Best is trial 27 with value: 0.6592308295470516.
bagging, val_score: 0.659231: 10%|# | 1/10 [00:00<00:06, 1.48it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004809 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.649764 valid's binary_logloss: 0.663649
[200] train's binary_logloss: 0.638988 valid's binary_logloss: 0.65972
[300] train's binary_logloss: 0.632186 valid's binary_logloss: 0.659366
[400] train's binary_logloss: 0.626354 valid's binary_logloss: 0.659639
Early stopping, best iteration is:
[322] train's binary_logloss: 0.630893 valid's binary_logloss: 0.659046
bagging, val_score: 0.659046: 20%|## | 2/10 [00:01<00:05, 1.42it/s][I 2020-09-27 04:48:24,511] Trial 28 finished with value: 0.659045827636086 and parameters: {'bagging_fraction': 0.8505997166142588, 'bagging_freq': 3}. Best is trial 28 with value: 0.659045827636086.
bagging, val_score: 0.659046: 20%|## | 2/10 [00:01<00:05, 1.42it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.006005 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448
[200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404
[300] train's binary_logloss: 0.631879 valid's binary_logloss: 0.657967
Early stopping, best iteration is:
[281] train's binary_logloss: 0.633137 valid's binary_logloss: 0.657825
bagging, val_score: 0.657825: 30%|### | 3/10 [00:02<00:05, 1.40it/s][I 2020-09-27 04:48:25,250] Trial 29 finished with value: 0.6578254229578112 and parameters: {'bagging_fraction': 0.8638496845155332, 'bagging_freq': 3}. Best is trial 29 with value: 0.6578254229578112.
bagging, val_score: 0.657825: 30%|### | 3/10 [00:02<00:05, 1.40it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001015 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.649823 valid's binary_logloss: 0.663663
[200] train's binary_logloss: 0.638792 valid's binary_logloss: 0.659288
[300] train's binary_logloss: 0.631889 valid's binary_logloss: 0.658487
Early stopping, best iteration is:
[241] train's binary_logloss: 0.635846 valid's binary_logloss: 0.658154
bagging, val_score: 0.657825: 40%|#### | 4/10 [00:03<00:04, 1.28it/s][I 2020-09-27 04:48:26,194] Trial 30 finished with value: 0.6581539755446953 and parameters: {'bagging_fraction': 0.8557672308630124, 'bagging_freq': 3}. Best is trial 29 with value: 0.6578254229578112.
bagging, val_score: 0.657825: 40%|#### | 4/10 [00:03<00:04, 1.28it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002755 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650081 valid's binary_logloss: 0.663971
[200] train's binary_logloss: 0.639132 valid's binary_logloss: 0.660066
[300] train's binary_logloss: 0.632409 valid's binary_logloss: 0.659247
[400] train's binary_logloss: 0.626618 valid's binary_logloss: 0.65931
Early stopping, best iteration is:
[341] train's binary_logloss: 0.630063 valid's binary_logloss: 0.659011
bagging, val_score: 0.657825: 50%|##### | 5/10 [00:04<00:04, 1.12it/s][I 2020-09-27 04:48:27,332] Trial 31 finished with value: 0.6590114175208794 and parameters: {'bagging_fraction': 0.8641735815144664, 'bagging_freq': 3}. Best is trial 29 with value: 0.6578254229578112.
bagging, val_score: 0.657825: 50%|##### | 5/10 [00:04<00:04, 1.12it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000894 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.64988 valid's binary_logloss: 0.664431
[200] train's binary_logloss: 0.639026 valid's binary_logloss: 0.660245
[300] train's binary_logloss: 0.632354 valid's binary_logloss: 0.659402
[400] train's binary_logloss: 0.626336 valid's binary_logloss: 0.659571
Early stopping, best iteration is:
[322] train's binary_logloss: 0.630965 valid's binary_logloss: 0.659137
bagging, val_score: 0.657825: 60%|###### | 6/10 [00:05<00:03, 1.17it/s][I 2020-09-27 04:48:28,113] Trial 32 finished with value: 0.6591365814377496 and parameters: {'bagging_fraction': 0.8683729064753997, 'bagging_freq': 3}. Best is trial 29 with value: 0.6578254229578112.
bagging, val_score: 0.657825: 60%|###### | 6/10 [00:05<00:03, 1.17it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.006198 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.649852 valid's binary_logloss: 0.663933
[200] train's binary_logloss: 0.639035 valid's binary_logloss: 0.65961
[300] train's binary_logloss: 0.632241 valid's binary_logloss: 0.658575
[400] train's binary_logloss: 0.62642 valid's binary_logloss: 0.658789
Early stopping, best iteration is:
[362] train's binary_logloss: 0.628634 valid's binary_logloss: 0.658088
bagging, val_score: 0.657825: 70%|####### | 7/10 [00:05<00:02, 1.16it/s][I 2020-09-27 04:48:28,978] Trial 33 finished with value: 0.6580884313848213 and parameters: {'bagging_fraction': 0.8714478568851503, 'bagging_freq': 3}. Best is trial 29 with value: 0.6578254229578112.
bagging, val_score: 0.657825: 70%|####### | 7/10 [00:05<00:02, 1.16it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000833 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.649854 valid's binary_logloss: 0.66428
[200] train's binary_logloss: 0.638784 valid's binary_logloss: 0.660031
[300] train's binary_logloss: 0.631851 valid's binary_logloss: 0.659733
Early stopping, best iteration is:
[259] train's binary_logloss: 0.634487 valid's binary_logloss: 0.659495
bagging, val_score: 0.657825: 80%|######## | 8/10 [00:06<00:01, 1.20it/s][I 2020-09-27 04:48:29,747] Trial 34 finished with value: 0.6594947433812736 and parameters: {'bagging_fraction': 0.9161669099599079, 'bagging_freq': 3}. Best is trial 29 with value: 0.6578254229578112.
bagging, val_score: 0.657825: 80%|######## | 8/10 [00:06<00:01, 1.20it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.003234 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.649122 valid's binary_logloss: 0.66312
[200] train's binary_logloss: 0.638562 valid's binary_logloss: 0.660896
Early stopping, best iteration is:
[170] train's binary_logloss: 0.640905 valid's binary_logloss: 0.660106
bagging, val_score: 0.657825: 90%|######### | 9/10 [00:07<00:00, 1.18it/s][I 2020-09-27 04:48:30,634] Trial 35 finished with value: 0.6601056363582504 and parameters: {'bagging_fraction': 0.5253666376896976, 'bagging_freq': 2}. Best is trial 29 with value: 0.6578254229578112.
bagging, val_score: 0.657825: 90%|######### | 9/10 [00:07<00:00, 1.18it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000924 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.64967 valid's binary_logloss: 0.663468
[200] train's binary_logloss: 0.638905 valid's binary_logloss: 0.660271
[300] train's binary_logloss: 0.632375 valid's binary_logloss: 0.660007
[400] train's binary_logloss: 0.626598 valid's binary_logloss: 0.659422
[500] train's binary_logloss: 0.621008 valid's binary_logloss: 0.658741
Early stopping, best iteration is:
[478] train's binary_logloss: 0.622298 valid's binary_logloss: 0.658486
bagging, val_score: 0.657825: 100%|##########| 10/10 [00:08<00:00, 1.12it/s][I 2020-09-27 04:48:31,633] Trial 36 finished with value: 0.658485788443909 and parameters: {'bagging_fraction': 0.7531472629337493, 'bagging_freq': 6}. Best is trial 29 with value: 0.6578254229578112.
bagging, val_score: 0.657825: 100%|##########| 10/10 [00:08<00:00, 1.17it/s]
feature_fraction_stage2, val_score: 0.657825: 0%| | 0/6 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000919 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.649945 valid's binary_logloss: 0.663854
[200] train's binary_logloss: 0.638845 valid's binary_logloss: 0.659895
[300] train's binary_logloss: 0.631902 valid's binary_logloss: 0.659391
[400] train's binary_logloss: 0.626123 valid's binary_logloss: 0.659153
Early stopping, best iteration is:
[373] train's binary_logloss: 0.627702 valid's binary_logloss: 0.658836
feature_fraction_stage2, val_score: 0.657825: 17%|#6 | 1/6 [00:00<00:04, 1.09it/s][I 2020-09-27 04:48:32,563] Trial 37 finished with value: 0.6588361247453068 and parameters: {'feature_fraction': 0.8480000000000001}. Best is trial 37 with value: 0.6588361247453068.
feature_fraction_stage2, val_score: 0.657825: 17%|#6 | 1/6 [00:00<00:04, 1.09it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000843 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448
[200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404
[300] train's binary_logloss: 0.631879 valid's binary_logloss: 0.657967
Early stopping, best iteration is:
[281] train's binary_logloss: 0.633137 valid's binary_logloss: 0.657825
feature_fraction_stage2, val_score: 0.657825: 33%|###3 | 2/6 [00:01<00:03, 1.16it/s][I 2020-09-27 04:48:33,305] Trial 38 finished with value: 0.6578254229578112 and parameters: {'feature_fraction': 0.8160000000000001}. Best is trial 38 with value: 0.6578254229578112.
feature_fraction_stage2, val_score: 0.657825: 33%|###3 | 2/6 [00:01<00:03, 1.16it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000906 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.649964 valid's binary_logloss: 0.664299
[200] train's binary_logloss: 0.638952 valid's binary_logloss: 0.66021
[300] train's binary_logloss: 0.632195 valid's binary_logloss: 0.6599
Early stopping, best iteration is:
[265] train's binary_logloss: 0.634377 valid's binary_logloss: 0.659183
feature_fraction_stage2, val_score: 0.657825: 50%|##### | 3/6 [00:02<00:02, 1.11it/s][I 2020-09-27 04:48:34,293] Trial 39 finished with value: 0.6591825632130718 and parameters: {'feature_fraction': 0.88}. Best is trial 38 with value: 0.6578254229578112.
feature_fraction_stage2, val_score: 0.657825: 50%|##### | 3/6 [00:02<00:02, 1.11it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016186 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650087 valid's binary_logloss: 0.664287
[200] train's binary_logloss: 0.639299 valid's binary_logloss: 0.66052
[300] train's binary_logloss: 0.632483 valid's binary_logloss: 0.659821
Early stopping, best iteration is:
[265] train's binary_logloss: 0.634723 valid's binary_logloss: 0.659476
feature_fraction_stage2, val_score: 0.657825: 67%|######6 | 4/6 [00:03<00:01, 1.12it/s][I 2020-09-27 04:48:35,160] Trial 40 finished with value: 0.6594764202643758 and parameters: {'feature_fraction': 0.784}. Best is trial 38 with value: 0.6578254229578112.
feature_fraction_stage2, val_score: 0.657825: 67%|######6 | 4/6 [00:03<00:01, 1.12it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005048 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650002 valid's binary_logloss: 0.663921
[200] train's binary_logloss: 0.639037 valid's binary_logloss: 0.659453
[300] train's binary_logloss: 0.632258 valid's binary_logloss: 0.658764
Early stopping, best iteration is:
[263] train's binary_logloss: 0.634669 valid's binary_logloss: 0.658316
feature_fraction_stage2, val_score: 0.657825: 83%|########3 | 5/6 [00:04<00:00, 1.21it/s][I 2020-09-27 04:48:35,831] Trial 41 finished with value: 0.6583158730564244 and parameters: {'feature_fraction': 0.7200000000000001}. Best is trial 38 with value: 0.6578254229578112.
feature_fraction_stage2, val_score: 0.657825: 83%|########3 | 5/6 [00:04<00:00, 1.21it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004831 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650087 valid's binary_logloss: 0.664287
[200] train's binary_logloss: 0.639299 valid's binary_logloss: 0.66052
[300] train's binary_logloss: 0.632483 valid's binary_logloss: 0.659821
Early stopping, best iteration is:
[265] train's binary_logloss: 0.634723 valid's binary_logloss: 0.659476
feature_fraction_stage2, val_score: 0.657825: 100%|##########| 6/6 [00:04<00:00, 1.29it/s][I 2020-09-27 04:48:36,497] Trial 42 finished with value: 0.6594764202643758 and parameters: {'feature_fraction': 0.7520000000000001}. Best is trial 38 with value: 0.6578254229578112.
feature_fraction_stage2, val_score: 0.657825: 100%|##########| 6/6 [00:04<00:00, 1.24it/s]
regularization_factors, val_score: 0.657825: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000810 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650995 valid's binary_logloss: 0.663984
[200] train's binary_logloss: 0.640816 valid's binary_logloss: 0.660341
[300] train's binary_logloss: 0.635324 valid's binary_logloss: 0.659307
Early stopping, best iteration is:
[279] train's binary_logloss: 0.636364 valid's binary_logloss: 0.659099
regularization_factors, val_score: 0.657825: 5%|5 | 1/20 [00:00<00:15, 1.24it/s][I 2020-09-27 04:48:37,323] Trial 43 finished with value: 0.659098571564232 and parameters: {'lambda_l1': 6.368584839611204, 'lambda_l2': 0.011340463712795283}. Best is trial 43 with value: 0.659098571564232.
regularization_factors, val_score: 0.657825: 5%|5 | 1/20 [00:00<00:15, 1.24it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004832 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448
[200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404
[300] train's binary_logloss: 0.631873 valid's binary_logloss: 0.657967
Early stopping, best iteration is:
[281] train's binary_logloss: 0.633131 valid's binary_logloss: 0.657825
regularization_factors, val_score: 0.657825: 10%|# | 2/20 [00:01<00:15, 1.13it/s][I 2020-09-27 04:48:38,394] Trial 44 finished with value: 0.6578254666437432 and parameters: {'lambda_l1': 2.7238144783184155e-08, 'lambda_l2': 1.008805152305263e-07}. Best is trial 44 with value: 0.6578254666437432.
regularization_factors, val_score: 0.657825: 10%|# | 2/20 [00:01<00:15, 1.13it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.021204 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448
[200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404
[300] train's binary_logloss: 0.631873 valid's binary_logloss: 0.657967
Early stopping, best iteration is:
[281] train's binary_logloss: 0.633131 valid's binary_logloss: 0.657825
regularization_factors, val_score: 0.657825: 15%|#5 | 3/20 [00:02<00:14, 1.13it/s][I 2020-09-27 04:48:39,260] Trial 45 finished with value: 0.6578254666410345 and parameters: {'lambda_l1': 1.0435559181670616e-08, 'lambda_l2': 1.585450495098862e-08}. Best is trial 45 with value: 0.6578254666410345.
regularization_factors, val_score: 0.657825: 15%|#5 | 3/20 [00:02<00:14, 1.13it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000902 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448
[200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404
[300] train's binary_logloss: 0.631876 valid's binary_logloss: 0.657967
Early stopping, best iteration is:
[281] train's binary_logloss: 0.633134 valid's binary_logloss: 0.657826
regularization_factors, val_score: 0.657825: 20%|## | 4/20 [00:03<00:13, 1.19it/s][I 2020-09-27 04:48:39,997] Trial 46 finished with value: 0.6578256353646856 and parameters: {'lambda_l1': 1.0219085602778263e-08, 'lambda_l2': 1.1221944608727081e-08}. Best is trial 45 with value: 0.6578254666410345.
regularization_factors, val_score: 0.657825: 20%|## | 4/20 [00:03<00:13, 1.19it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001003 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448
[200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404
[300] train's binary_logloss: 0.631876 valid's binary_logloss: 0.657967
Early stopping, best iteration is:
[281] train's binary_logloss: 0.633134 valid's binary_logloss: 0.657826
regularization_factors, val_score: 0.657825: 25%|##5 | 5/20 [00:04<00:12, 1.22it/s][I 2020-09-27 04:48:40,774] Trial 47 finished with value: 0.6578256353644696 and parameters: {'lambda_l1': 1.0248936810635019e-08, 'lambda_l2': 1.3858509908150066e-08}. Best is trial 45 with value: 0.6578254666410345.
regularization_factors, val_score: 0.657825: 25%|##5 | 5/20 [00:04<00:12, 1.22it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000912 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448
[200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404
[300] train's binary_logloss: 0.631876 valid's binary_logloss: 0.657967
Early stopping, best iteration is:
[281] train's binary_logloss: 0.633134 valid's binary_logloss: 0.657826
regularization_factors, val_score: 0.657825: 30%|### | 6/20 [00:04<00:11, 1.26it/s][I 2020-09-27 04:48:41,504] Trial 48 finished with value: 0.6578256353646091 and parameters: {'lambda_l1': 1.4330977464180962e-08, 'lambda_l2': 1.3975871503118737e-08}. Best is trial 45 with value: 0.6578254666410345.
regularization_factors, val_score: 0.657825: 30%|### | 6/20 [00:04<00:11, 1.26it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008196 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448
[200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404
[300] train's binary_logloss: 0.631873 valid's binary_logloss: 0.657967
Early stopping, best iteration is:
[281] train's binary_logloss: 0.633131 valid's binary_logloss: 0.657825
regularization_factors, val_score: 0.657825: 35%|###5 | 7/20 [00:06<00:11, 1.10it/s][I 2020-09-27 04:48:42,673] Trial 49 finished with value: 0.6578254666417223 and parameters: {'lambda_l1': 1.3796608447685168e-08, 'lambda_l2': 4.7462022137544425e-08}. Best is trial 45 with value: 0.6578254666410345.
regularization_factors, val_score: 0.657825: 35%|###5 | 7/20 [00:06<00:11, 1.10it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000848 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448
[200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404
[300] train's binary_logloss: 0.631873 valid's binary_logloss: 0.657967
Early stopping, best iteration is:
[281] train's binary_logloss: 0.633131 valid's binary_logloss: 0.657825
regularization_factors, val_score: 0.657825: 40%|#### | 8/20 [00:07<00:10, 1.13it/s][I 2020-09-27 04:48:43,519] Trial 50 finished with value: 0.6578254844386293 and parameters: {'lambda_l1': 5.179213600763938e-06, 'lambda_l2': 5.7935679948530245e-06}. Best is trial 45 with value: 0.6578254666410345.
regularization_factors, val_score: 0.657825: 40%|#### | 8/20 [00:07<00:10, 1.13it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000955 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448
[200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404
[300] train's binary_logloss: 0.631876 valid's binary_logloss: 0.657967
Early stopping, best iteration is:
[281] train's binary_logloss: 0.633134 valid's binary_logloss: 0.657825
regularization_factors, val_score: 0.657825: 45%|####5 | 9/20 [00:07<00:09, 1.19it/s][I 2020-09-27 04:48:44,255] Trial 51 finished with value: 0.6578252553356203 and parameters: {'lambda_l1': 1.558360532559241e-05, 'lambda_l2': 4.619266691410038e-06}. Best is trial 51 with value: 0.6578252553356203.
regularization_factors, val_score: 0.657825: 45%|####5 | 9/20 [00:07<00:09, 1.19it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000894 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448
[200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404
[300] train's binary_logloss: 0.631876 valid's binary_logloss: 0.657967
Early stopping, best iteration is:
[281] train's binary_logloss: 0.633134 valid's binary_logloss: 0.657826
regularization_factors, val_score: 0.657825: 50%|##### | 10/20 [00:08<00:08, 1.23it/s][I 2020-09-27 04:48:44,994] Trial 52 finished with value: 0.6578256356080326 and parameters: {'lambda_l1': 2.7250062441770353e-06, 'lambda_l2': 2.367717105214214e-06}. Best is trial 51 with value: 0.6578252553356203.
regularization_factors, val_score: 0.657825: 50%|##### | 10/20 [00:08<00:08, 1.23it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001566 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448
[200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404
[300] train's binary_logloss: 0.631876 valid's binary_logloss: 0.657967
Early stopping, best iteration is:
[281] train's binary_logloss: 0.633134 valid's binary_logloss: 0.657826
regularization_factors, val_score: 0.657825: 55%|#####5 | 11/20 [00:09<00:07, 1.26it/s][I 2020-09-27 04:48:45,738] Trial 53 finished with value: 0.6578256354185161 and parameters: {'lambda_l1': 3.3319990687032807e-07, 'lambda_l2': 9.864870800228018e-07}. Best is trial 51 with value: 0.6578252553356203.
regularization_factors, val_score: 0.657825: 55%|#####5 | 11/20 [00:09<00:07, 1.26it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000843 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650018 valid's binary_logloss: 0.663448
[200] train's binary_logloss: 0.638999 valid's binary_logloss: 0.659122
[300] train's binary_logloss: 0.632026 valid's binary_logloss: 0.658235
Early stopping, best iteration is:
[290] train's binary_logloss: 0.632696 valid's binary_logloss: 0.658147
regularization_factors, val_score: 0.657825: 60%|###### | 12/20 [00:10<00:07, 1.08it/s][I 2020-09-27 04:48:46,984] Trial 54 finished with value: 0.6581469442698293 and parameters: {'lambda_l1': 0.006687572450441869, 'lambda_l2': 2.42706687165855e-07}. Best is trial 51 with value: 0.6578252553356203.
regularization_factors, val_score: 0.657825: 60%|###### | 12/20 [00:10<00:07, 1.08it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000800 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448
[200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404
[300] train's binary_logloss: 0.631879 valid's binary_logloss: 0.657967
Early stopping, best iteration is:
[281] train's binary_logloss: 0.633137 valid's binary_logloss: 0.657825
regularization_factors, val_score: 0.657825: 65%|######5 | 13/20 [00:11<00:06, 1.14it/s][I 2020-09-27 04:48:47,745] Trial 55 finished with value: 0.6578254639325656 and parameters: {'lambda_l1': 0.0006287455216203462, 'lambda_l2': 8.873368461516288e-05}. Best is trial 51 with value: 0.6578252553356203.
regularization_factors, val_score: 0.657825: 65%|######5 | 13/20 [00:11<00:06, 1.14it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000955 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448
[200] train's binary_logloss: 0.638998 valid's binary_logloss: 0.659122
[300] train's binary_logloss: 0.632035 valid's binary_logloss: 0.658192
[400] train's binary_logloss: 0.626308 valid's binary_logloss: 0.65838
Early stopping, best iteration is:
[307] train's binary_logloss: 0.631536 valid's binary_logloss: 0.658173
regularization_factors, val_score: 0.657825: 70%|####### | 14/20 [00:12<00:05, 1.18it/s][I 2020-09-27 04:48:48,519] Trial 56 finished with value: 0.6581734662791846 and parameters: {'lambda_l1': 0.0023930535028393245, 'lambda_l2': 0.00017393958564666502}. Best is trial 51 with value: 0.6578252553356203.
regularization_factors, val_score: 0.657825: 70%|####### | 14/20 [00:12<00:05, 1.18it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000968 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448
[200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404
[300] train's binary_logloss: 0.631879 valid's binary_logloss: 0.657967
Early stopping, best iteration is:
[281] train's binary_logloss: 0.633137 valid's binary_logloss: 0.657825
regularization_factors, val_score: 0.657825: 75%|#######5 | 15/20 [00:12<00:04, 1.23it/s][I 2020-09-27 04:48:49,262] Trial 57 finished with value: 0.6578254362526145 and parameters: {'lambda_l1': 0.0001139415321377858, 'lambda_l2': 0.00018316916066158292}. Best is trial 51 with value: 0.6578252553356203.
regularization_factors, val_score: 0.657825: 75%|#######5 | 15/20 [00:12<00:04, 1.23it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001132 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448
[200] train's binary_logloss: 0.63892 valid's binary_logloss: 0.659404
[300] train's binary_logloss: 0.63188 valid's binary_logloss: 0.657967
Early stopping, best iteration is:
[281] train's binary_logloss: 0.633137 valid's binary_logloss: 0.657825
regularization_factors, val_score: 0.657825: 80%|######## | 16/20 [00:13<00:03, 1.16it/s][I 2020-09-27 04:48:50,236] Trial 58 finished with value: 0.6578254556273145 and parameters: {'lambda_l1': 0.00013528915615576415, 'lambda_l2': 0.0002063652161436333}. Best is trial 51 with value: 0.6578252553356203.
regularization_factors, val_score: 0.657825: 80%|######## | 16/20 [00:13<00:03, 1.16it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.015940 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448
[200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404
[300] train's binary_logloss: 0.631876 valid's binary_logloss: 0.657967
Early stopping, best iteration is:
[281] train's binary_logloss: 0.633134 valid's binary_logloss: 0.657826
regularization_factors, val_score: 0.657825: 85%|########5 | 17/20 [00:14<00:02, 1.13it/s][I 2020-09-27 04:48:51,159] Trial 59 finished with value: 0.6578256499831218 and parameters: {'lambda_l1': 0.00013602675042342853, 'lambda_l2': 0.00018327214045940748}. Best is trial 51 with value: 0.6578252553356203.
regularization_factors, val_score: 0.657825: 85%|########5 | 17/20 [00:14<00:02, 1.13it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001556 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448
[200] train's binary_logloss: 0.63892 valid's binary_logloss: 0.659405
[300] train's binary_logloss: 0.631879 valid's binary_logloss: 0.657967
Early stopping, best iteration is:
[281] train's binary_logloss: 0.633136 valid's binary_logloss: 0.657826
regularization_factors, val_score: 0.657825: 90%|######### | 18/20 [00:15<00:01, 1.20it/s][I 2020-09-27 04:48:51,873] Trial 60 finished with value: 0.6578258958216437 and parameters: {'lambda_l1': 0.00015253198331570547, 'lambda_l2': 0.0071388436435203435}. Best is trial 51 with value: 0.6578252553356203.
regularization_factors, val_score: 0.657825: 90%|######### | 18/20 [00:15<00:01, 1.20it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000941 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448
[200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404
[300] train's binary_logloss: 0.63188 valid's binary_logloss: 0.657967
Early stopping, best iteration is:
[281] train's binary_logloss: 0.633137 valid's binary_logloss: 0.657826
regularization_factors, val_score: 0.657825: 95%|#########5| 19/20 [00:16<00:00, 1.23it/s][I 2020-09-27 04:48:52,641] Trial 61 finished with value: 0.657825517008465 and parameters: {'lambda_l1': 0.0015382833109009928, 'lambda_l2': 4.130282202202608e-05}. Best is trial 51 with value: 0.6578252553356203.
regularization_factors, val_score: 0.657825: 95%|#########5| 19/20 [00:16<00:00, 1.23it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000998 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448
[200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404
[300] train's binary_logloss: 0.631874 valid's binary_logloss: 0.657967
Early stopping, best iteration is:
[281] train's binary_logloss: 0.633131 valid's binary_logloss: 0.657826
regularization_factors, val_score: 0.657825: 100%|##########| 20/20 [00:16<00:00, 1.24it/s][I 2020-09-27 04:48:53,434] Trial 62 finished with value: 0.6578255316540529 and parameters: {'lambda_l1': 3.339088375221171e-05, 'lambda_l2': 0.0017874187322570777}. Best is trial 51 with value: 0.6578252553356203.
regularization_factors, val_score: 0.657825: 100%|##########| 20/20 [00:16<00:00, 1.18it/s]
min_data_in_leaf, val_score: 0.657825: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000868 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448
[200] train's binary_logloss: 0.638922 valid's binary_logloss: 0.659362
[300] train's binary_logloss: 0.632151 valid's binary_logloss: 0.658294
[400] train's binary_logloss: 0.626333 valid's binary_logloss: 0.659009
Early stopping, best iteration is:
[300] train's binary_logloss: 0.632151 valid's binary_logloss: 0.658294
min_data_in_leaf, val_score: 0.657825: 20%|## | 1/5 [00:01<00:05, 1.28s/it][I 2020-09-27 04:48:54,726] Trial 63 finished with value: 0.6582938352043944 and parameters: {'min_child_samples': 25}. Best is trial 63 with value: 0.6582938352043944.
min_data_in_leaf, val_score: 0.657825: 20%|## | 1/5 [00:01<00:05, 1.28s/it][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005115 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448
[200] train's binary_logloss: 0.638826 valid's binary_logloss: 0.659409
[300] train's binary_logloss: 0.63179 valid's binary_logloss: 0.658711
Early stopping, best iteration is:
[232] train's binary_logloss: 0.636374 valid's binary_logloss: 0.658604
min_data_in_leaf, val_score: 0.657825: 40%|#### | 2/5 [00:01<00:03, 1.10s/it][I 2020-09-27 04:48:55,420] Trial 64 finished with value: 0.6586044923878769 and parameters: {'min_child_samples': 5}. Best is trial 63 with value: 0.6582938352043944.
min_data_in_leaf, val_score: 0.657825: 40%|#### | 2/5 [00:01<00:03, 1.10s/it][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000876 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.65002 valid's binary_logloss: 0.663404
[200] train's binary_logloss: 0.639166 valid's binary_logloss: 0.659615
[300] train's binary_logloss: 0.632623 valid's binary_logloss: 0.659448
Early stopping, best iteration is:
[240] train's binary_logloss: 0.636379 valid's binary_logloss: 0.65892
min_data_in_leaf, val_score: 0.657825: 60%|###### | 3/5 [00:02<00:01, 1.02it/s][I 2020-09-27 04:48:56,123] Trial 65 finished with value: 0.658919688761251 and parameters: {'min_child_samples': 50}. Best is trial 63 with value: 0.6582938352043944.
min_data_in_leaf, val_score: 0.657825: 60%|###### | 3/5 [00:02<00:01, 1.02it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004999 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.65002 valid's binary_logloss: 0.663404
[200] train's binary_logloss: 0.639125 valid's binary_logloss: 0.659535
[300] train's binary_logloss: 0.632949 valid's binary_logloss: 0.658458
Early stopping, best iteration is:
[266] train's binary_logloss: 0.634844 valid's binary_logloss: 0.658133
min_data_in_leaf, val_score: 0.657825: 80%|######## | 4/5 [00:03<00:00, 1.10it/s][I 2020-09-27 04:48:56,850] Trial 66 finished with value: 0.6581334712449208 and parameters: {'min_child_samples': 100}. Best is trial 66 with value: 0.6581334712449208.
min_data_in_leaf, val_score: 0.657825: 80%|######## | 4/5 [00:03<00:00, 1.10it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000869 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4241
[LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001
[LightGBM] [Info] Start training from score -0.024001
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448
[200] train's binary_logloss: 0.638826 valid's binary_logloss: 0.659409
[300] train's binary_logloss: 0.631757 valid's binary_logloss: 0.658712
Early stopping, best iteration is:
[234] train's binary_logloss: 0.636229 valid's binary_logloss: 0.65847
min_data_in_leaf, val_score: 0.657825: 100%|##########| 5/5 [00:04<00:00, 1.19it/s][I 2020-09-27 04:48:57,537] Trial 67 finished with value: 0.6584702414733883 and parameters: {'min_child_samples': 10}. Best is trial 66 with value: 0.6581334712449208.
min_data_in_leaf, val_score: 0.657825: 100%|##########| 5/5 [00:04<00:00, 1.22it/s]
################################
CV_score:0.6150646280811062
Fold : 0
[I 2020-09-27 04:48:57,616] A new study created in memory with name: no-name-8ff89ddf-7641-453c-86d0-fc507cfacb11
feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.026169 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663649 valid's binary_logloss: 0.690213
Early stopping, best iteration is:
[45] train's binary_logloss: 0.676671 valid's binary_logloss: 0.689632
feature_fraction, val_score: 0.689632: 14%|#4 | 1/7 [00:01<00:08, 1.45s/it][I 2020-09-27 04:48:59,075] Trial 0 finished with value: 0.6896322280348909 and parameters: {'feature_fraction': 0.7}. Best is trial 0 with value: 0.6896322280348909.
feature_fraction, val_score: 0.689632: 14%|#4 | 1/7 [00:01<00:08, 1.45s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001410 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.66617 valid's binary_logloss: 0.690438
Early stopping, best iteration is:
[41] train's binary_logloss: 0.679097 valid's binary_logloss: 0.689783
feature_fraction, val_score: 0.689632: 29%|##8 | 2/7 [00:02<00:06, 1.24s/it][I 2020-09-27 04:48:59,837] Trial 1 finished with value: 0.6897826459094011 and parameters: {'feature_fraction': 0.4}. Best is trial 0 with value: 0.6896322280348909.
feature_fraction, val_score: 0.689632: 29%|##8 | 2/7 [00:02<00:06, 1.24s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.003154 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.662081 valid's binary_logloss: 0.690003
Early stopping, best iteration is:
[41] train's binary_logloss: 0.677046 valid's binary_logloss: 0.68968
feature_fraction, val_score: 0.689632: 43%|####2 | 3/7 [00:03<00:04, 1.13s/it][I 2020-09-27 04:49:00,694] Trial 2 finished with value: 0.6896804566917485 and parameters: {'feature_fraction': 1.0}. Best is trial 0 with value: 0.6896322280348909.
feature_fraction, val_score: 0.689632: 43%|####2 | 3/7 [00:03<00:04, 1.13s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.015461 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.66415 valid's binary_logloss: 0.689593
Early stopping, best iteration is:
[39] train's binary_logloss: 0.678462 valid's binary_logloss: 0.689021
feature_fraction, val_score: 0.689021: 57%|#####7 | 4/7 [00:03<00:03, 1.01s/it][I 2020-09-27 04:49:01,437] Trial 3 finished with value: 0.6890207233896298 and parameters: {'feature_fraction': 0.6}. Best is trial 3 with value: 0.6890207233896298.
feature_fraction, val_score: 0.689021: 57%|#####7 | 4/7 [00:03<00:03, 1.01s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.025191 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.66251 valid's binary_logloss: 0.689598
Early stopping, best iteration is:
[42] train's binary_logloss: 0.676825 valid's binary_logloss: 0.68942
feature_fraction, val_score: 0.689021: 71%|#######1 | 5/7 [00:05<00:02, 1.13s/it][I 2020-09-27 04:49:02,837] Trial 4 finished with value: 0.6894197474355058 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 3 with value: 0.6890207233896298.
feature_fraction, val_score: 0.689021: 71%|#######1 | 5/7 [00:05<00:02, 1.13s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.017648 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663408 valid's binary_logloss: 0.690473
Early stopping, best iteration is:
[45] train's binary_logloss: 0.676529 valid's binary_logloss: 0.689828
feature_fraction, val_score: 0.689021: 86%|########5 | 6/7 [00:06<00:01, 1.04s/it][I 2020-09-27 04:49:03,685] Trial 5 finished with value: 0.68982789311769 and parameters: {'feature_fraction': 0.8}. Best is trial 3 with value: 0.6890207233896298.
feature_fraction, val_score: 0.689021: 86%|########5 | 6/7 [00:06<00:01, 1.04s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001730 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665183 valid's binary_logloss: 0.689631
[200] train's binary_logloss: 0.645808 valid's binary_logloss: 0.690591
Early stopping, best iteration is:
[102] train's binary_logloss: 0.664779 valid's binary_logloss: 0.689546
feature_fraction, val_score: 0.689021: 100%|##########| 7/7 [00:07<00:00, 1.07s/it][I 2020-09-27 04:49:04,813] Trial 6 finished with value: 0.6895458054912846 and parameters: {'feature_fraction': 0.5}. Best is trial 3 with value: 0.6890207233896298.
feature_fraction, val_score: 0.689021: 100%|##########| 7/7 [00:07<00:00, 1.03s/it]
num_leaves, val_score: 0.689021: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012666 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.577872 valid's binary_logloss: 0.693632
Early stopping, best iteration is:
[18] train's binary_logloss: 0.663683 valid's binary_logloss: 0.689631
num_leaves, val_score: 0.689021: 5%|5 | 1/20 [00:01<00:30, 1.61s/it][I 2020-09-27 04:49:06,444] Trial 7 finished with value: 0.689631097968691 and parameters: {'num_leaves': 168}. Best is trial 7 with value: 0.689631097968691.
num_leaves, val_score: 0.689021: 5%|5 | 1/20 [00:01<00:30, 1.61s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.017419 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.604202 valid's binary_logloss: 0.69167
Early stopping, best iteration is:
[38] train's binary_logloss: 0.650476 valid's binary_logloss: 0.689618
num_leaves, val_score: 0.689021: 10%|# | 2/20 [00:02<00:26, 1.47s/it][I 2020-09-27 04:49:07,566] Trial 8 finished with value: 0.689618324134334 and parameters: {'num_leaves': 122}. Best is trial 8 with value: 0.689618324134334.
num_leaves, val_score: 0.689021: 10%|# | 2/20 [00:02<00:26, 1.47s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016515 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685172 valid's binary_logloss: 0.689666
[200] train's binary_logloss: 0.681255 valid's binary_logloss: 0.689678
Early stopping, best iteration is:
[156] train's binary_logloss: 0.68289 valid's binary_logloss: 0.689534
num_leaves, val_score: 0.689021: 15%|#5 | 3/20 [00:03<00:22, 1.34s/it][I 2020-09-27 04:49:08,626] Trial 9 finished with value: 0.6895342400114649 and parameters: {'num_leaves': 6}. Best is trial 9 with value: 0.6895342400114649.
num_leaves, val_score: 0.689021: 15%|#5 | 3/20 [00:03<00:22, 1.34s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016558 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.535492 valid's binary_logloss: 0.695541
Early stopping, best iteration is:
[19] train's binary_logloss: 0.649759 valid's binary_logloss: 0.690794
num_leaves, val_score: 0.689021: 20%|## | 4/20 [00:05<00:23, 1.49s/it][I 2020-09-27 04:49:10,461] Trial 10 finished with value: 0.6907940845828584 and parameters: {'num_leaves': 253}. Best is trial 9 with value: 0.6895342400114649.
num_leaves, val_score: 0.689021: 20%|## | 4/20 [00:05<00:23, 1.49s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.017448 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.67101 valid's binary_logloss: 0.689552
Early stopping, best iteration is:
[93] train's binary_logloss: 0.672156 valid's binary_logloss: 0.68942
num_leaves, val_score: 0.689021: 25%|##5 | 5/20 [00:06<00:19, 1.31s/it][I 2020-09-27 04:49:11,362] Trial 11 finished with value: 0.6894197097930296 and parameters: {'num_leaves': 22}. Best is trial 11 with value: 0.6894197097930296.
num_leaves, val_score: 0.689021: 25%|##5 | 5/20 [00:06<00:19, 1.31s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.014207 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.690044 valid's binary_logloss: 0.690612
[200] train's binary_logloss: 0.68885 valid's binary_logloss: 0.689962
[300] train's binary_logloss: 0.688098 valid's binary_logloss: 0.689612
[400] train's binary_logloss: 0.687569 valid's binary_logloss: 0.689487
[500] train's binary_logloss: 0.68718 valid's binary_logloss: 0.689391
[600] train's binary_logloss: 0.686881 valid's binary_logloss: 0.689318
[700] train's binary_logloss: 0.68664 valid's binary_logloss: 0.689265
[800] train's binary_logloss: 0.68644 valid's binary_logloss: 0.689264
Early stopping, best iteration is:
[710] train's binary_logloss: 0.686618 valid's binary_logloss: 0.689244
num_leaves, val_score: 0.689021: 30%|### | 6/20 [00:09<00:25, 1.86s/it][I 2020-09-27 04:49:14,484] Trial 12 finished with value: 0.689244475322615 and parameters: {'num_leaves': 2}. Best is trial 12 with value: 0.689244475322615.
num_leaves, val_score: 0.689021: 30%|### | 6/20 [00:09<00:25, 1.86s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.015182 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.630223 valid's binary_logloss: 0.692152
Early stopping, best iteration is:
[21] train's binary_logloss: 0.674331 valid's binary_logloss: 0.690049
num_leaves, val_score: 0.689021: 35%|###5 | 7/20 [00:10<00:20, 1.56s/it][I 2020-09-27 04:49:15,354] Trial 13 finished with value: 0.6900491534448284 and parameters: {'num_leaves': 80}. Best is trial 12 with value: 0.689244475322615.
num_leaves, val_score: 0.689021: 35%|###5 | 7/20 [00:10<00:20, 1.56s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.017787 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.543727 valid's binary_logloss: 0.696815
Early stopping, best iteration is:
[10] train's binary_logloss: 0.670334 valid's binary_logloss: 0.691482
num_leaves, val_score: 0.689021: 40%|#### | 8/20 [00:11<00:17, 1.46s/it][I 2020-09-27 04:49:16,588] Trial 14 finished with value: 0.6914823998645343 and parameters: {'num_leaves': 234}. Best is trial 12 with value: 0.689244475322615.
num_leaves, val_score: 0.689021: 40%|#### | 8/20 [00:11<00:17, 1.46s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.015457 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.646446 valid's binary_logloss: 0.691884
Early stopping, best iteration is:
[17] train's binary_logloss: 0.681433 valid's binary_logloss: 0.689734
num_leaves, val_score: 0.689021: 45%|####5 | 9/20 [00:13<00:15, 1.43s/it][I 2020-09-27 04:49:17,939] Trial 15 finished with value: 0.6897344485270362 and parameters: {'num_leaves': 55}. Best is trial 12 with value: 0.689244475322615.
num_leaves, val_score: 0.689021: 45%|####5 | 9/20 [00:13<00:15, 1.43s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.015412 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.570477 valid's binary_logloss: 0.695696
Early stopping, best iteration is:
[23] train's binary_logloss: 0.654169 valid's binary_logloss: 0.690909
num_leaves, val_score: 0.689021: 50%|##### | 10/20 [00:14<00:13, 1.34s/it][I 2020-09-27 04:49:19,067] Trial 16 finished with value: 0.6909092950799796 and parameters: {'num_leaves': 181}. Best is trial 12 with value: 0.689244475322615.
num_leaves, val_score: 0.689021: 50%|##### | 10/20 [00:14<00:13, 1.34s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016435 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.623443 valid's binary_logloss: 0.691768
Early stopping, best iteration is:
[19] train's binary_logloss: 0.674343 valid's binary_logloss: 0.690276
num_leaves, val_score: 0.689021: 55%|#####5 | 11/20 [00:15<00:10, 1.20s/it][I 2020-09-27 04:49:19,941] Trial 17 finished with value: 0.690276336147683 and parameters: {'num_leaves': 90}. Best is trial 12 with value: 0.689244475322615.
num_leaves, val_score: 0.689021: 55%|#####5 | 11/20 [00:15<00:10, 1.20s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016181 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.659158 valid's binary_logloss: 0.689655
Early stopping, best iteration is:
[35] train's binary_logloss: 0.677405 valid's binary_logloss: 0.689049
num_leaves, val_score: 0.689021: 60%|###### | 12/20 [00:15<00:08, 1.07s/it][I 2020-09-27 04:49:20,717] Trial 18 finished with value: 0.6890491443530252 and parameters: {'num_leaves': 38}. Best is trial 18 with value: 0.6890491443530252.
num_leaves, val_score: 0.689021: 60%|###### | 12/20 [00:15<00:08, 1.07s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.023330 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.652371 valid's binary_logloss: 0.690333
Early stopping, best iteration is:
[39] train's binary_logloss: 0.672949 valid's binary_logloss: 0.68919
num_leaves, val_score: 0.689021: 65%|######5 | 13/20 [00:17<00:08, 1.17s/it][I 2020-09-27 04:49:22,101] Trial 19 finished with value: 0.6891896975164848 and parameters: {'num_leaves': 47}. Best is trial 18 with value: 0.6890491443530252.
num_leaves, val_score: 0.689021: 65%|######5 | 13/20 [00:17<00:08, 1.17s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.017299 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.59964 valid's binary_logloss: 0.69374
Early stopping, best iteration is:
[14] train's binary_logloss: 0.674012 valid's binary_logloss: 0.690104
num_leaves, val_score: 0.689021: 70%|####### | 14/20 [00:18<00:06, 1.10s/it][I 2020-09-27 04:49:23,062] Trial 20 finished with value: 0.6901038597516463 and parameters: {'num_leaves': 130}. Best is trial 18 with value: 0.6890491443530252.
num_leaves, val_score: 0.689021: 70%|####### | 14/20 [00:18<00:06, 1.10s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016669 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657163 valid's binary_logloss: 0.690788
Early stopping, best iteration is:
[43] train's binary_logloss: 0.674068 valid's binary_logloss: 0.689599
num_leaves, val_score: 0.689021: 75%|#######5 | 15/20 [00:19<00:05, 1.01s/it][I 2020-09-27 04:49:23,851] Trial 21 finished with value: 0.6895994157631041 and parameters: {'num_leaves': 40}. Best is trial 18 with value: 0.6890491443530252.
num_leaves, val_score: 0.689021: 75%|#######5 | 15/20 [00:19<00:05, 1.01s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016387 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.632347 valid's binary_logloss: 0.692318
Early stopping, best iteration is:
[23] train's binary_logloss: 0.673649 valid's binary_logloss: 0.689645
num_leaves, val_score: 0.689021: 80%|######## | 16/20 [00:19<00:03, 1.05it/s][I 2020-09-27 04:49:24,677] Trial 22 finished with value: 0.6896448463652016 and parameters: {'num_leaves': 76}. Best is trial 18 with value: 0.6890491443530252.
num_leaves, val_score: 0.689021: 80%|######## | 16/20 [00:19<00:03, 1.05it/s][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.020197 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656137 valid's binary_logloss: 0.689636
Early stopping, best iteration is:
[96] train's binary_logloss: 0.657272 valid's binary_logloss: 0.689437
num_leaves, val_score: 0.689021: 85%|########5 | 17/20 [00:21<00:03, 1.10s/it][I 2020-09-27 04:49:26,120] Trial 23 finished with value: 0.6894373899731695 and parameters: {'num_leaves': 42}. Best is trial 18 with value: 0.6890491443530252.
num_leaves, val_score: 0.689021: 85%|########5 | 17/20 [00:21<00:03, 1.10s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016258 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.608119 valid's binary_logloss: 0.691911
Early stopping, best iteration is:
[30] train's binary_logloss: 0.659223 valid's binary_logloss: 0.689435
num_leaves, val_score: 0.689021: 90%|######### | 18/20 [00:22<00:02, 1.09s/it][I 2020-09-27 04:49:27,187] Trial 24 finished with value: 0.6894354954462946 and parameters: {'num_leaves': 116}. Best is trial 18 with value: 0.6890491443530252.
num_leaves, val_score: 0.689021: 90%|######### | 18/20 [00:22<00:02, 1.09s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.018058 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68615 valid's binary_logloss: 0.689604
[200] train's binary_logloss: 0.682919 valid's binary_logloss: 0.689378
Early stopping, best iteration is:
[158] train's binary_logloss: 0.684169 valid's binary_logloss: 0.689219
num_leaves, val_score: 0.689021: 95%|#########5| 19/20 [00:23<00:01, 1.08s/it][I 2020-09-27 04:49:28,242] Trial 25 finished with value: 0.6892189395196165 and parameters: {'num_leaves': 5}. Best is trial 18 with value: 0.6890491443530252.
num_leaves, val_score: 0.689021: 95%|#########5| 19/20 [00:23<00:01, 1.08s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016101 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.651145 valid's binary_logloss: 0.690351
Early stopping, best iteration is:
[34] train's binary_logloss: 0.674412 valid's binary_logloss: 0.689378
num_leaves, val_score: 0.689021: 100%|##########| 20/20 [00:24<00:00, 1.16s/it][I 2020-09-27 04:49:29,591] Trial 26 finished with value: 0.6893783362619839 and parameters: {'num_leaves': 49}. Best is trial 18 with value: 0.6890491443530252.
num_leaves, val_score: 0.689021: 100%|##########| 20/20 [00:24<00:00, 1.24s/it]
bagging, val_score: 0.689021: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013092 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664277 valid's binary_logloss: 0.68939
Early stopping, best iteration is:
[71] train's binary_logloss: 0.670651 valid's binary_logloss: 0.688825
bagging, val_score: 0.688825: 10%|# | 1/10 [00:01<00:09, 1.08s/it][I 2020-09-27 04:49:30,680] Trial 27 finished with value: 0.6888247862401992 and parameters: {'bagging_fraction': 0.958245225148382, 'bagging_freq': 2}. Best is trial 27 with value: 0.6888247862401992.
bagging, val_score: 0.688825: 10%|# | 1/10 [00:01<00:09, 1.08s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.014041 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664031 valid's binary_logloss: 0.689087
[200] train's binary_logloss: 0.64388 valid's binary_logloss: 0.689768
Early stopping, best iteration is:
[119] train's binary_logloss: 0.659997 valid's binary_logloss: 0.688879
bagging, val_score: 0.688825: 20%|## | 2/10 [00:02<00:09, 1.15s/it][I 2020-09-27 04:49:32,003] Trial 28 finished with value: 0.6888794888870088 and parameters: {'bagging_fraction': 0.9745012448452529, 'bagging_freq': 2}. Best is trial 27 with value: 0.6888247862401992.
bagging, val_score: 0.688825: 20%|## | 2/10 [00:02<00:09, 1.15s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.015566 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664149 valid's binary_logloss: 0.689861
Early stopping, best iteration is:
[50] train's binary_logloss: 0.675702 valid's binary_logloss: 0.689193
bagging, val_score: 0.688825: 30%|### | 3/10 [00:03<00:08, 1.23s/it][I 2020-09-27 04:49:33,428] Trial 29 finished with value: 0.689193182327793 and parameters: {'bagging_fraction': 0.9667819838074765, 'bagging_freq': 2}. Best is trial 27 with value: 0.6888247862401992.
bagging, val_score: 0.688825: 30%|### | 3/10 [00:03<00:08, 1.23s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013341 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664296 valid's binary_logloss: 0.689547
Early stopping, best iteration is:
[60] train's binary_logloss: 0.673228 valid's binary_logloss: 0.689074
bagging, val_score: 0.688825: 40%|#### | 4/10 [00:04<00:07, 1.17s/it][I 2020-09-27 04:49:34,451] Trial 30 finished with value: 0.6890740654309894 and parameters: {'bagging_fraction': 0.9690126597280045, 'bagging_freq': 2}. Best is trial 27 with value: 0.6888247862401992.
bagging, val_score: 0.688825: 40%|#### | 4/10 [00:04<00:07, 1.17s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016944 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664297 valid's binary_logloss: 0.690412
Early stopping, best iteration is:
[54] train's binary_logloss: 0.674733 valid's binary_logloss: 0.68958
bagging, val_score: 0.688825: 50%|##### | 5/10 [00:05<00:05, 1.10s/it][I 2020-09-27 04:49:35,398] Trial 31 finished with value: 0.6895802218638041 and parameters: {'bagging_fraction': 0.9943804516504096, 'bagging_freq': 4}. Best is trial 27 with value: 0.6888247862401992.
bagging, val_score: 0.688825: 50%|##### | 5/10 [00:05<00:05, 1.10s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.014817 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664664 valid's binary_logloss: 0.689957
Early stopping, best iteration is:
[37] train's binary_logloss: 0.679397 valid's binary_logloss: 0.689326
bagging, val_score: 0.688825: 60%|###### | 6/10 [00:06<00:04, 1.07s/it][I 2020-09-27 04:49:36,386] Trial 32 finished with value: 0.6893263473853064 and parameters: {'bagging_fraction': 0.6476701297352184, 'bagging_freq': 1}. Best is trial 27 with value: 0.6888247862401992.
bagging, val_score: 0.688825: 60%|###### | 6/10 [00:06<00:04, 1.07s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.029663 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515
Early stopping, best iteration is:
[49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719
bagging, val_score: 0.688719: 70%|####### | 7/10 [00:08<00:03, 1.16s/it][I 2020-09-27 04:49:37,766] Trial 33 finished with value: 0.6887188408517487 and parameters: {'bagging_fraction': 0.8287757616049252, 'bagging_freq': 7}. Best is trial 33 with value: 0.6887188408517487.
bagging, val_score: 0.688719: 70%|####### | 7/10 [00:08<00:03, 1.16s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012644 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.66436 valid's binary_logloss: 0.689716
Early stopping, best iteration is:
[51] train's binary_logloss: 0.675497 valid's binary_logloss: 0.689039
bagging, val_score: 0.688719: 80%|######## | 8/10 [00:09<00:02, 1.09s/it][I 2020-09-27 04:49:38,691] Trial 34 finished with value: 0.689039279019024 and parameters: {'bagging_fraction': 0.8282760453593162, 'bagging_freq': 7}. Best is trial 33 with value: 0.6887188408517487.
bagging, val_score: 0.688719: 80%|######## | 8/10 [00:09<00:02, 1.09s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016881 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663999 valid's binary_logloss: 0.690093
Early stopping, best iteration is:
[92] train's binary_logloss: 0.665835 valid's binary_logloss: 0.689992
bagging, val_score: 0.688719: 90%|######### | 9/10 [00:10<00:01, 1.10s/it][I 2020-09-27 04:49:39,803] Trial 35 finished with value: 0.6899923113836022 and parameters: {'bagging_fraction': 0.8214825248241318, 'bagging_freq': 4}. Best is trial 33 with value: 0.6887188408517487.
bagging, val_score: 0.688719: 90%|######### | 9/10 [00:10<00:01, 1.10s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013803 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664239 valid's binary_logloss: 0.688911
[200] train's binary_logloss: 0.644408 valid's binary_logloss: 0.690305
Early stopping, best iteration is:
[103] train's binary_logloss: 0.663554 valid's binary_logloss: 0.688799
bagging, val_score: 0.688719: 100%|##########| 10/10 [00:11<00:00, 1.24s/it][I 2020-09-27 04:49:41,362] Trial 36 finished with value: 0.6887991231168928 and parameters: {'bagging_fraction': 0.8383907363062589, 'bagging_freq': 7}. Best is trial 33 with value: 0.6887188408517487.
bagging, val_score: 0.688719: 100%|##########| 10/10 [00:11<00:00, 1.18s/it]
feature_fraction_stage2, val_score: 0.688719: 0%| | 0/6 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013665 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515
Early stopping, best iteration is:
[49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719
feature_fraction_stage2, val_score: 0.688719: 17%|#6 | 1/6 [00:00<00:04, 1.03it/s][I 2020-09-27 04:49:42,349] Trial 37 finished with value: 0.6887188408517487 and parameters: {'feature_fraction': 0.616}. Best is trial 37 with value: 0.6887188408517487.
feature_fraction_stage2, val_score: 0.688719: 17%|#6 | 1/6 [00:00<00:04, 1.03it/s][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001729 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664799 valid's binary_logloss: 0.689634
Early stopping, best iteration is:
[69] train's binary_logloss: 0.671673 valid's binary_logloss: 0.689345
feature_fraction_stage2, val_score: 0.688719: 33%|###3 | 2/6 [00:02<00:03, 1.00it/s][I 2020-09-27 04:49:43,417] Trial 38 finished with value: 0.6893450614153805 and parameters: {'feature_fraction': 0.552}. Best is trial 37 with value: 0.6887188408517487.
feature_fraction_stage2, val_score: 0.688719: 33%|###3 | 2/6 [00:02<00:03, 1.00it/s][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001868 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664799 valid's binary_logloss: 0.689634
Early stopping, best iteration is:
[69] train's binary_logloss: 0.671673 valid's binary_logloss: 0.689345
feature_fraction_stage2, val_score: 0.688719: 50%|##### | 3/6 [00:03<00:03, 1.17s/it][I 2020-09-27 04:49:44,990] Trial 39 finished with value: 0.6893450614153805 and parameters: {'feature_fraction': 0.52}. Best is trial 37 with value: 0.6887188408517487.
feature_fraction_stage2, val_score: 0.688719: 50%|##### | 3/6 [00:03<00:03, 1.17s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001814 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664285 valid's binary_logloss: 0.690298
Early stopping, best iteration is:
[64] train's binary_logloss: 0.672325 valid's binary_logloss: 0.689397
feature_fraction_stage2, val_score: 0.688719: 67%|######6 | 4/6 [00:04<00:02, 1.14s/it][I 2020-09-27 04:49:46,067] Trial 40 finished with value: 0.6893965384793862 and parameters: {'feature_fraction': 0.584}. Best is trial 37 with value: 0.6887188408517487.
feature_fraction_stage2, val_score: 0.688719: 67%|######6 | 4/6 [00:04<00:02, 1.14s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013983 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663794 valid's binary_logloss: 0.689901
[200] train's binary_logloss: 0.64382 valid's binary_logloss: 0.690825
Early stopping, best iteration is:
[115] train's binary_logloss: 0.66065 valid's binary_logloss: 0.689797
feature_fraction_stage2, val_score: 0.688719: 83%|########3 | 5/6 [00:06<00:01, 1.42s/it][I 2020-09-27 04:49:48,122] Trial 41 finished with value: 0.6897973746145477 and parameters: {'feature_fraction': 0.6799999999999999}. Best is trial 37 with value: 0.6887188408517487.
feature_fraction_stage2, val_score: 0.688719: 83%|########3 | 5/6 [00:06<00:01, 1.42s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.031387 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663746 valid's binary_logloss: 0.689073
[200] train's binary_logloss: 0.643757 valid's binary_logloss: 0.691217
Early stopping, best iteration is:
[105] train's binary_logloss: 0.662622 valid's binary_logloss: 0.689054
feature_fraction_stage2, val_score: 0.688719: 100%|##########| 6/6 [00:08<00:00, 1.44s/it][I 2020-09-27 04:49:49,607] Trial 42 finished with value: 0.6890541064139101 and parameters: {'feature_fraction': 0.6479999999999999}. Best is trial 37 with value: 0.6887188408517487.
feature_fraction_stage2, val_score: 0.688719: 100%|##########| 6/6 [00:08<00:00, 1.37s/it]
regularization_factors, val_score: 0.688719: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012472 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664985 valid's binary_logloss: 0.690223
[200] train's binary_logloss: 0.646149 valid's binary_logloss: 0.690617
Early stopping, best iteration is:
[151] train's binary_logloss: 0.655125 valid's binary_logloss: 0.690101
regularization_factors, val_score: 0.688719: 5%|5 | 1/20 [00:01<00:28, 1.50s/it][I 2020-09-27 04:49:51,126] Trial 43 finished with value: 0.6901011755065565 and parameters: {'lambda_l1': 1.1303470188424833, 'lambda_l2': 0.06375127708462985}. Best is trial 43 with value: 0.6901011755065565.
regularization_factors, val_score: 0.688719: 5%|5 | 1/20 [00:01<00:28, 1.50s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.014406 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515
Early stopping, best iteration is:
[49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719
regularization_factors, val_score: 0.688719: 10%|# | 2/20 [00:02<00:26, 1.47s/it][I 2020-09-27 04:49:52,538] Trial 44 finished with value: 0.6887188408527825 and parameters: {'lambda_l1': 1.8636898004930976e-08, 'lambda_l2': 6.192880162509011e-08}. Best is trial 44 with value: 0.6887188408527825.
regularization_factors, val_score: 0.688719: 10%|# | 2/20 [00:02<00:26, 1.47s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016047 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515
Early stopping, best iteration is:
[49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719
regularization_factors, val_score: 0.688719: 15%|#5 | 3/20 [00:03<00:22, 1.32s/it][I 2020-09-27 04:49:53,496] Trial 45 finished with value: 0.6887188408522336 and parameters: {'lambda_l1': 1.8439298966850396e-08, 'lambda_l2': 1.3511479384741456e-08}. Best is trial 45 with value: 0.6887188408522336.
regularization_factors, val_score: 0.688719: 15%|#5 | 3/20 [00:03<00:22, 1.32s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012081 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515
Early stopping, best iteration is:
[49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719
regularization_factors, val_score: 0.688719: 20%|## | 4/20 [00:04<00:19, 1.20s/it][I 2020-09-27 04:49:54,422] Trial 46 finished with value: 0.6887188408519217 and parameters: {'lambda_l1': 1.2530425311422802e-08, 'lambda_l2': 1.0756585467257088e-08}. Best is trial 46 with value: 0.6887188408519217.
regularization_factors, val_score: 0.688719: 20%|## | 4/20 [00:04<00:19, 1.20s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.014442 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515
Early stopping, best iteration is:
[49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719
regularization_factors, val_score: 0.688719: 25%|##5 | 5/20 [00:05<00:16, 1.13s/it][I 2020-09-27 04:49:55,370] Trial 47 finished with value: 0.6887188408519394 and parameters: {'lambda_l1': 1.2148997058183217e-08, 'lambda_l2': 1.076670603152709e-08}. Best is trial 46 with value: 0.6887188408519217.
regularization_factors, val_score: 0.688719: 25%|##5 | 5/20 [00:05<00:16, 1.13s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.015325 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515
Early stopping, best iteration is:
[49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719
regularization_factors, val_score: 0.688719: 30%|### | 6/20 [00:07<00:16, 1.21s/it][I 2020-09-27 04:49:56,764] Trial 48 finished with value: 0.6887188408519016 and parameters: {'lambda_l1': 1.0473302644937364e-08, 'lambda_l2': 1.4000255193861959e-08}. Best is trial 48 with value: 0.6887188408519016.
regularization_factors, val_score: 0.688719: 30%|### | 6/20 [00:07<00:16, 1.21s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.017485 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515
Early stopping, best iteration is:
[49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719
regularization_factors, val_score: 0.688719: 35%|###5 | 7/20 [00:08<00:14, 1.15s/it][I 2020-09-27 04:49:57,778] Trial 49 finished with value: 0.6887188408520548 and parameters: {'lambda_l1': 1.2933235345738932e-08, 'lambda_l2': 1.3210044579433778e-08}. Best is trial 48 with value: 0.6887188408519016.
regularization_factors, val_score: 0.688719: 35%|###5 | 7/20 [00:08<00:14, 1.15s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016188 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515
Early stopping, best iteration is:
[49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719
regularization_factors, val_score: 0.688719: 40%|#### | 8/20 [00:09<00:13, 1.09s/it][I 2020-09-27 04:49:58,722] Trial 50 finished with value: 0.6887188410169663 and parameters: {'lambda_l1': 6.522515649363731e-06, 'lambda_l2': 5.399603910938485e-06}. Best is trial 48 with value: 0.6887188408519016.
regularization_factors, val_score: 0.688719: 40%|#### | 8/20 [00:09<00:13, 1.09s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.018393 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515
Early stopping, best iteration is:
[49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719
regularization_factors, val_score: 0.688719: 45%|####5 | 9/20 [00:10<00:12, 1.18s/it][I 2020-09-27 04:50:00,120] Trial 51 finished with value: 0.6887188408522066 and parameters: {'lambda_l1': 1.7781053835321766e-08, 'lambda_l2': 1.2911207447672734e-08}. Best is trial 48 with value: 0.6887188408519016.
regularization_factors, val_score: 0.688719: 45%|####5 | 9/20 [00:10<00:12, 1.18s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012528 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515
Early stopping, best iteration is:
[49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719
regularization_factors, val_score: 0.688719: 50%|##### | 10/20 [00:11<00:11, 1.14s/it][I 2020-09-27 04:50:01,154] Trial 52 finished with value: 0.6887188408521625 and parameters: {'lambda_l1': 1.487838960944784e-08, 'lambda_l2': 1.2171741979090901e-08}. Best is trial 48 with value: 0.6887188408519016.
regularization_factors, val_score: 0.688719: 50%|##### | 10/20 [00:11<00:11, 1.14s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013835 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515
Early stopping, best iteration is:
[49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719
regularization_factors, val_score: 0.688719: 55%|#####5 | 11/20 [00:12<00:09, 1.07s/it][I 2020-09-27 04:50:02,068] Trial 53 finished with value: 0.688718840880323 and parameters: {'lambda_l1': 1.2436063977852807e-06, 'lambda_l2': 5.416531879194707e-07}. Best is trial 48 with value: 0.6887188408519016.
regularization_factors, val_score: 0.688719: 55%|#####5 | 11/20 [00:12<00:09, 1.07s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.020591 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515
Early stopping, best iteration is:
[49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719
regularization_factors, val_score: 0.688719: 60%|###### | 12/20 [00:13<00:08, 1.03s/it][I 2020-09-27 04:50:03,019] Trial 54 finished with value: 0.6887188408518048 and parameters: {'lambda_l1': 1.001927452586009e-08, 'lambda_l2': 1.0173335682514728e-08}. Best is trial 54 with value: 0.6887188408518048.
regularization_factors, val_score: 0.688719: 60%|###### | 12/20 [00:13<00:08, 1.03s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013735 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515
Early stopping, best iteration is:
[49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719
regularization_factors, val_score: 0.688719: 65%|######5 | 13/20 [00:14<00:07, 1.13s/it][I 2020-09-27 04:50:04,380] Trial 55 finished with value: 0.6887188408925199 and parameters: {'lambda_l1': 1.0246196926378488e-06, 'lambda_l2': 2.2586132878897466e-06}. Best is trial 54 with value: 0.6887188408518048.
regularization_factors, val_score: 0.688719: 65%|######5 | 13/20 [00:14<00:07, 1.13s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.017187 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664481 valid's binary_logloss: 0.689645
Early stopping, best iteration is:
[49] train's binary_logloss: 0.675961 valid's binary_logloss: 0.689284
regularization_factors, val_score: 0.688719: 70%|####### | 14/20 [00:15<00:06, 1.08s/it][I 2020-09-27 04:50:05,336] Trial 56 finished with value: 0.6892843278626516 and parameters: {'lambda_l1': 0.04091667144851904, 'lambda_l2': 1.0750374226083052e-08}. Best is trial 54 with value: 0.6887188408518048.
regularization_factors, val_score: 0.688719: 70%|####### | 14/20 [00:15<00:06, 1.08s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013760 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515
Early stopping, best iteration is:
[49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719
regularization_factors, val_score: 0.688719: 75%|#######5 | 15/20 [00:16<00:05, 1.04s/it][I 2020-09-27 04:50:06,270] Trial 57 finished with value: 0.6887188408561253 and parameters: {'lambda_l1': 1.5233181813455253e-07, 'lambda_l2': 1.8743684852533287e-07}. Best is trial 54 with value: 0.6887188408518048.
regularization_factors, val_score: 0.688719: 75%|#######5 | 15/20 [00:16<00:05, 1.04s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.015130 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515
Early stopping, best iteration is:
[49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719
regularization_factors, val_score: 0.688719: 80%|######## | 16/20 [00:19<00:06, 1.55s/it][I 2020-09-27 04:50:09,004] Trial 58 finished with value: 0.6887188422885628 and parameters: {'lambda_l1': 1.744760689790351e-07, 'lambda_l2': 0.00013088268252619333}. Best is trial 54 with value: 0.6887188408518048.
regularization_factors, val_score: 0.688719: 80%|######## | 16/20 [00:19<00:06, 1.55s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.018473 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664768 valid's binary_logloss: 0.689166
Early stopping, best iteration is:
[85] train's binary_logloss: 0.667917 valid's binary_logloss: 0.688911
regularization_factors, val_score: 0.688719: 85%|########5 | 17/20 [00:20<00:04, 1.45s/it][I 2020-09-27 04:50:10,235] Trial 59 finished with value: 0.6889111184760159 and parameters: {'lambda_l1': 0.00016154773417025348, 'lambda_l2': 0.9174839426871738}. Best is trial 54 with value: 0.6887188408518048.
regularization_factors, val_score: 0.688719: 85%|########5 | 17/20 [00:20<00:04, 1.45s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.015601 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515
Early stopping, best iteration is:
[49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719
regularization_factors, val_score: 0.688719: 90%|######### | 18/20 [00:21<00:02, 1.32s/it][I 2020-09-27 04:50:11,235] Trial 60 finished with value: 0.6887188408571195 and parameters: {'lambda_l1': 1.0344224061447487e-07, 'lambda_l2': 3.115824250207294e-07}. Best is trial 54 with value: 0.6887188408518048.
regularization_factors, val_score: 0.688719: 90%|######### | 18/20 [00:21<00:02, 1.32s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016453 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515
Early stopping, best iteration is:
[49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719
regularization_factors, val_score: 0.688719: 95%|#########5| 19/20 [00:23<00:01, 1.34s/it][I 2020-09-27 04:50:12,647] Trial 61 finished with value: 0.6887188408520247 and parameters: {'lambda_l1': 1.3726047963300931e-08, 'lambda_l2': 1.1171463229442362e-08}. Best is trial 54 with value: 0.6887188408518048.
regularization_factors, val_score: 0.688719: 95%|#########5| 19/20 [00:23<00:01, 1.34s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.003196 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515
Early stopping, best iteration is:
[49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719
regularization_factors, val_score: 0.688719: 100%|##########| 20/20 [00:24<00:00, 1.26s/it][I 2020-09-27 04:50:13,695] Trial 62 finished with value: 0.688718840851812 and parameters: {'lambda_l1': 1.0200361074017904e-08, 'lambda_l2': 1.1298563130408035e-08}. Best is trial 54 with value: 0.6887188408518048.
regularization_factors, val_score: 0.688719: 100%|##########| 20/20 [00:24<00:00, 1.20s/it]
min_data_in_leaf, val_score: 0.688719: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013241 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664243 valid's binary_logloss: 0.689415
Early stopping, best iteration is:
[56] train's binary_logloss: 0.674277 valid's binary_logloss: 0.689398
min_data_in_leaf, val_score: 0.688719: 20%|## | 1/5 [00:00<00:03, 1.06it/s][I 2020-09-27 04:50:14,651] Trial 63 finished with value: 0.6893981290180272 and parameters: {'min_child_samples': 25}. Best is trial 63 with value: 0.6893981290180272.
min_data_in_leaf, val_score: 0.688719: 20%|## | 1/5 [00:00<00:03, 1.06it/s][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012878 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663871 valid's binary_logloss: 0.689525
[200] train's binary_logloss: 0.643739 valid's binary_logloss: 0.690786
Early stopping, best iteration is:
[124] train's binary_logloss: 0.658769 valid's binary_logloss: 0.689265
min_data_in_leaf, val_score: 0.688719: 40%|#### | 2/5 [00:02<00:03, 1.17s/it][I 2020-09-27 04:50:16,344] Trial 64 finished with value: 0.6892650557557729 and parameters: {'min_child_samples': 5}. Best is trial 64 with value: 0.6892650557557729.
min_data_in_leaf, val_score: 0.688719: 40%|#### | 2/5 [00:02<00:03, 1.17s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013201 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664176 valid's binary_logloss: 0.689954
Early stopping, best iteration is:
[55] train's binary_logloss: 0.674447 valid's binary_logloss: 0.689227
min_data_in_leaf, val_score: 0.688719: 60%|###### | 3/5 [00:03<00:02, 1.14s/it][I 2020-09-27 04:50:17,422] Trial 65 finished with value: 0.6892266404784428 and parameters: {'min_child_samples': 10}. Best is trial 65 with value: 0.6892266404784428.
min_data_in_leaf, val_score: 0.688719: 60%|###### | 3/5 [00:03<00:02, 1.14s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013886 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664683 valid's binary_logloss: 0.689526
Early stopping, best iteration is:
[69] train's binary_logloss: 0.67151 valid's binary_logloss: 0.689297
min_data_in_leaf, val_score: 0.688719: 80%|######## | 4/5 [00:04<00:01, 1.13s/it][I 2020-09-27 04:50:18,524] Trial 66 finished with value: 0.689296784041159 and parameters: {'min_child_samples': 50}. Best is trial 65 with value: 0.6892266404784428.
min_data_in_leaf, val_score: 0.688719: 80%|######## | 4/5 [00:04<00:01, 1.13s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013526 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106
[LightGBM] [Info] Start training from score 0.004106
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665727 valid's binary_logloss: 0.689486
Early stopping, best iteration is:
[74] train's binary_logloss: 0.67097 valid's binary_logloss: 0.689201
min_data_in_leaf, val_score: 0.688719: 100%|##########| 5/5 [00:06<00:00, 1.19s/it][I 2020-09-27 04:50:19,869] Trial 67 finished with value: 0.689200711916551 and parameters: {'min_child_samples': 100}. Best is trial 67 with value: 0.689200711916551.
min_data_in_leaf, val_score: 0.688719: 100%|##########| 5/5 [00:06<00:00, 1.24s/it]
Fold : 1
[I 2020-09-27 04:50:19,990] A new study created in memory with name: no-name-42491456-2d3a-4d80-ab38-ce544f64ff2c
feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001692 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665024 valid's binary_logloss: 0.689794
Early stopping, best iteration is:
[70] train's binary_logloss: 0.671608 valid's binary_logloss: 0.689532
feature_fraction, val_score: 0.689532: 14%|#4 | 1/7 [00:01<00:07, 1.22s/it][I 2020-09-27 04:50:21,234] Trial 0 finished with value: 0.689531715864851 and parameters: {'feature_fraction': 0.5}. Best is trial 0 with value: 0.689531715864851.
feature_fraction, val_score: 0.689532: 14%|#4 | 1/7 [00:01<00:07, 1.22s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012878 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663919 valid's binary_logloss: 0.690116
Early stopping, best iteration is:
[40] train's binary_logloss: 0.678175 valid's binary_logloss: 0.689468
feature_fraction, val_score: 0.689468: 29%|##8 | 2/7 [00:01<00:05, 1.09s/it][I 2020-09-27 04:50:22,004] Trial 1 finished with value: 0.6894679163318443 and parameters: {'feature_fraction': 0.6}. Best is trial 1 with value: 0.6894679163318443.
feature_fraction, val_score: 0.689468: 29%|##8 | 2/7 [00:01<00:05, 1.09s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011687 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.66358 valid's binary_logloss: 0.690887
Early stopping, best iteration is:
[35] train's binary_logloss: 0.679357 valid's binary_logloss: 0.690079
feature_fraction, val_score: 0.689468: 43%|####2 | 3/7 [00:02<00:03, 1.01it/s][I 2020-09-27 04:50:22,771] Trial 2 finished with value: 0.6900787550751419 and parameters: {'feature_fraction': 0.7}. Best is trial 1 with value: 0.6894679163318443.
feature_fraction, val_score: 0.689468: 43%|####2 | 3/7 [00:02<00:03, 1.01it/s][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.003338 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.662262 valid's binary_logloss: 0.690303
Early stopping, best iteration is:
[93] train's binary_logloss: 0.664122 valid's binary_logloss: 0.689979
feature_fraction, val_score: 0.689468: 57%|#####7 | 4/7 [00:04<00:03, 1.18s/it][I 2020-09-27 04:50:24,375] Trial 3 finished with value: 0.6899793275720499 and parameters: {'feature_fraction': 1.0}. Best is trial 1 with value: 0.6894679163318443.
feature_fraction, val_score: 0.689468: 57%|#####7 | 4/7 [00:04<00:03, 1.18s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.021087 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.662515 valid's binary_logloss: 0.689981
Early stopping, best iteration is:
[51] train's binary_logloss: 0.674455 valid's binary_logloss: 0.689748
feature_fraction, val_score: 0.689468: 71%|#######1 | 5/7 [00:05<00:02, 1.10s/it][I 2020-09-27 04:50:25,314] Trial 4 finished with value: 0.6897477331954484 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 1 with value: 0.6894679163318443.
feature_fraction, val_score: 0.689468: 71%|#######1 | 5/7 [00:05<00:02, 1.10s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001352 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.666503 valid's binary_logloss: 0.689198
[200] train's binary_logloss: 0.647913 valid's binary_logloss: 0.690111
Early stopping, best iteration is:
[141] train's binary_logloss: 0.658411 valid's binary_logloss: 0.689014
feature_fraction, val_score: 0.689014: 86%|########5 | 6/7 [00:06<00:01, 1.12s/it][I 2020-09-27 04:50:26,470] Trial 5 finished with value: 0.6890136743093267 and parameters: {'feature_fraction': 0.4}. Best is trial 5 with value: 0.6890136743093267.
feature_fraction, val_score: 0.689014: 86%|########5 | 6/7 [00:06<00:01, 1.12s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.018804 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.66328 valid's binary_logloss: 0.690498
Early stopping, best iteration is:
[29] train's binary_logloss: 0.681042 valid's binary_logloss: 0.689914
feature_fraction, val_score: 0.689014: 100%|##########| 7/7 [00:07<00:00, 1.02s/it][I 2020-09-27 04:50:27,266] Trial 6 finished with value: 0.6899136884834198 and parameters: {'feature_fraction': 0.8}. Best is trial 5 with value: 0.6890136743093267.
feature_fraction, val_score: 0.689014: 100%|##########| 7/7 [00:07<00:00, 1.04s/it]
num_leaves, val_score: 0.689014: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001398 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.59502 valid's binary_logloss: 0.693529
Early stopping, best iteration is:
[25] train's binary_logloss: 0.660259 valid's binary_logloss: 0.690335
num_leaves, val_score: 0.689014: 5%|5 | 1/20 [00:01<00:30, 1.59s/it][I 2020-09-27 04:50:28,875] Trial 7 finished with value: 0.6903350331651174 and parameters: {'num_leaves': 152}. Best is trial 7 with value: 0.6903350331651174.
num_leaves, val_score: 0.689014: 5%|5 | 1/20 [00:01<00:30, 1.59s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001113 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.621859 valid's binary_logloss: 0.693386
Early stopping, best iteration is:
[24] train's binary_logloss: 0.669901 valid's binary_logloss: 0.690519
num_leaves, val_score: 0.689014: 10%|# | 2/20 [00:02<00:25, 1.40s/it][I 2020-09-27 04:50:29,827] Trial 8 finished with value: 0.6905190321264499 and parameters: {'num_leaves': 103}. Best is trial 7 with value: 0.6903350331651174.
num_leaves, val_score: 0.689014: 10%|# | 2/20 [00:02<00:25, 1.40s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001303 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.607653 valid's binary_logloss: 0.691023
Early stopping, best iteration is:
[31] train's binary_logloss: 0.658737 valid's binary_logloss: 0.690547
num_leaves, val_score: 0.689014: 15%|#5 | 3/20 [00:03<00:22, 1.30s/it][I 2020-09-27 04:50:30,899] Trial 9 finished with value: 0.6905470890370582 and parameters: {'num_leaves': 128}. Best is trial 7 with value: 0.6903350331651174.
num_leaves, val_score: 0.689014: 15%|#5 | 3/20 [00:03<00:22, 1.30s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001257 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68554 valid's binary_logloss: 0.689548
[200] train's binary_logloss: 0.681807 valid's binary_logloss: 0.689004
[300] train's binary_logloss: 0.678526 valid's binary_logloss: 0.688996
Early stopping, best iteration is:
[221] train's binary_logloss: 0.681102 valid's binary_logloss: 0.688886
num_leaves, val_score: 0.688886: 20%|## | 4/20 [00:05<00:23, 1.44s/it][I 2020-09-27 04:50:32,672] Trial 10 finished with value: 0.6888861673616007 and parameters: {'num_leaves': 6}. Best is trial 10 with value: 0.6888861673616007.
num_leaves, val_score: 0.688886: 20%|## | 4/20 [00:05<00:23, 1.44s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001441 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.674935 valid's binary_logloss: 0.689151
[200] train's binary_logloss: 0.663169 valid's binary_logloss: 0.689994
Early stopping, best iteration is:
[122] train's binary_logloss: 0.672063 valid's binary_logloss: 0.689042
num_leaves, val_score: 0.688886: 25%|##5 | 5/20 [00:06<00:19, 1.33s/it][I 2020-09-27 04:50:33,728] Trial 11 finished with value: 0.6890424580437142 and parameters: {'num_leaves': 19}. Best is trial 10 with value: 0.6888861673616007.
num_leaves, val_score: 0.688886: 25%|##5 | 5/20 [00:06<00:19, 1.33s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001371 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688532 valid's binary_logloss: 0.690109
[200] train's binary_logloss: 0.686793 valid's binary_logloss: 0.689219
[300] train's binary_logloss: 0.685512 valid's binary_logloss: 0.68876
[400] train's binary_logloss: 0.684466 valid's binary_logloss: 0.688694
[500] train's binary_logloss: 0.68348 valid's binary_logloss: 0.6886
Early stopping, best iteration is:
[438] train's binary_logloss: 0.684084 valid's binary_logloss: 0.68855
num_leaves, val_score: 0.688550: 30%|### | 6/20 [00:08<00:23, 1.66s/it][I 2020-09-27 04:50:36,153] Trial 12 finished with value: 0.6885499863950102 and parameters: {'num_leaves': 3}. Best is trial 12 with value: 0.6885499863950102.
num_leaves, val_score: 0.688550: 30%|### | 6/20 [00:08<00:23, 1.66s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001342 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68554 valid's binary_logloss: 0.689548
[200] train's binary_logloss: 0.681807 valid's binary_logloss: 0.689004
[300] train's binary_logloss: 0.678526 valid's binary_logloss: 0.688996
Early stopping, best iteration is:
[221] train's binary_logloss: 0.681102 valid's binary_logloss: 0.688886
num_leaves, val_score: 0.688550: 35%|###5 | 7/20 [00:10<00:20, 1.54s/it][I 2020-09-27 04:50:37,431] Trial 13 finished with value: 0.6888861673616007 and parameters: {'num_leaves': 6}. Best is trial 12 with value: 0.6885499863950102.
num_leaves, val_score: 0.688550: 35%|###5 | 7/20 [00:10<00:20, 1.54s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000758 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.657781 valid's binary_logloss: 0.68976
Early stopping, best iteration is:
[90] train's binary_logloss: 0.660403 valid's binary_logloss: 0.6895
num_leaves, val_score: 0.688550: 40%|#### | 8/20 [00:11<00:16, 1.36s/it][I 2020-09-27 04:50:38,375] Trial 14 finished with value: 0.6895000449518462 and parameters: {'num_leaves': 43}. Best is trial 12 with value: 0.6885499863950102.
num_leaves, val_score: 0.688550: 40%|#### | 8/20 [00:11<00:16, 1.36s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000790 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.64493 valid's binary_logloss: 0.691503
Early stopping, best iteration is:
[44] train's binary_logloss: 0.667465 valid's binary_logloss: 0.690163
num_leaves, val_score: 0.688550: 45%|####5 | 9/20 [00:11<00:13, 1.20s/it][I 2020-09-27 04:50:39,181] Trial 15 finished with value: 0.6901631360080164 and parameters: {'num_leaves': 64}. Best is trial 12 with value: 0.6885499863950102.
num_leaves, val_score: 0.688550: 45%|####5 | 9/20 [00:11<00:13, 1.20s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000739 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.548818 valid's binary_logloss: 0.697085
Early stopping, best iteration is:
[20] train's binary_logloss: 0.653082 valid's binary_logloss: 0.690873
num_leaves, val_score: 0.688550: 50%|##### | 10/20 [00:13<00:12, 1.22s/it][I 2020-09-27 04:50:40,456] Trial 16 finished with value: 0.6908727065690785 and parameters: {'num_leaves': 247}. Best is trial 12 with value: 0.6885499863950102.
num_leaves, val_score: 0.688550: 50%|##### | 10/20 [00:13<00:12, 1.22s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000719 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.690028 valid's binary_logloss: 0.690957
[200] train's binary_logloss: 0.688879 valid's binary_logloss: 0.690086
[300] train's binary_logloss: 0.688169 valid's binary_logloss: 0.689501
[400] train's binary_logloss: 0.687668 valid's binary_logloss: 0.689075
[500] train's binary_logloss: 0.687288 valid's binary_logloss: 0.688845
[600] train's binary_logloss: 0.686991 valid's binary_logloss: 0.688707
[700] train's binary_logloss: 0.68675 valid's binary_logloss: 0.688619
[800] train's binary_logloss: 0.686547 valid's binary_logloss: 0.688562
[900] train's binary_logloss: 0.686373 valid's binary_logloss: 0.688526
[1000] train's binary_logloss: 0.686221 valid's binary_logloss: 0.688477
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.686221 valid's binary_logloss: 0.688477
num_leaves, val_score: 0.688477: 55%|#####5 | 11/20 [00:16<00:15, 1.75s/it][I 2020-09-27 04:50:43,436] Trial 17 finished with value: 0.6884769144658212 and parameters: {'num_leaves': 2}. Best is trial 17 with value: 0.6884769144658212.
num_leaves, val_score: 0.688477: 55%|#####5 | 11/20 [00:16<00:15, 1.75s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000850 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.573685 valid's binary_logloss: 0.696971
Early stopping, best iteration is:
[20] train's binary_logloss: 0.660038 valid's binary_logloss: 0.691352
num_leaves, val_score: 0.688477: 60%|###### | 12/20 [00:17<00:12, 1.57s/it][I 2020-09-27 04:50:44,593] Trial 18 finished with value: 0.6913522035050282 and parameters: {'num_leaves': 195}. Best is trial 17 with value: 0.6884769144658212.
num_leaves, val_score: 0.688477: 60%|###### | 12/20 [00:17<00:12, 1.57s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000815 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.64326 valid's binary_logloss: 0.690631
Early stopping, best iteration is:
[71] train's binary_logloss: 0.654351 valid's binary_logloss: 0.690055
num_leaves, val_score: 0.688477: 65%|######5 | 13/20 [00:18<00:09, 1.39s/it][I 2020-09-27 04:50:45,549] Trial 19 finished with value: 0.6900553800103405 and parameters: {'num_leaves': 66}. Best is trial 17 with value: 0.6884769144658212.
num_leaves, val_score: 0.688477: 65%|######5 | 13/20 [00:18<00:09, 1.39s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000857 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68554 valid's binary_logloss: 0.689548
[200] train's binary_logloss: 0.681807 valid's binary_logloss: 0.689004
[300] train's binary_logloss: 0.678526 valid's binary_logloss: 0.688996
Early stopping, best iteration is:
[221] train's binary_logloss: 0.681102 valid's binary_logloss: 0.688886
num_leaves, val_score: 0.688477: 70%|####### | 14/20 [00:19<00:07, 1.30s/it][I 2020-09-27 04:50:46,632] Trial 20 finished with value: 0.6888861673616007 and parameters: {'num_leaves': 6}. Best is trial 17 with value: 0.6884769144658212.
num_leaves, val_score: 0.688477: 70%|####### | 14/20 [00:19<00:07, 1.30s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000852 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688532 valid's binary_logloss: 0.690109
[200] train's binary_logloss: 0.686793 valid's binary_logloss: 0.689219
[300] train's binary_logloss: 0.685512 valid's binary_logloss: 0.68876
[400] train's binary_logloss: 0.684466 valid's binary_logloss: 0.688694
[500] train's binary_logloss: 0.68348 valid's binary_logloss: 0.6886
Early stopping, best iteration is:
[438] train's binary_logloss: 0.684084 valid's binary_logloss: 0.68855
num_leaves, val_score: 0.688477: 75%|#######5 | 15/20 [00:21<00:07, 1.42s/it][I 2020-09-27 04:50:48,340] Trial 21 finished with value: 0.6885499863950102 and parameters: {'num_leaves': 3}. Best is trial 17 with value: 0.6884769144658212.
num_leaves, val_score: 0.688477: 75%|#######5 | 15/20 [00:21<00:07, 1.42s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000875 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.660592 valid's binary_logloss: 0.690578
Early stopping, best iteration is:
[41] train's binary_logloss: 0.676299 valid's binary_logloss: 0.690019
num_leaves, val_score: 0.688477: 80%|######## | 16/20 [00:21<00:04, 1.22s/it][I 2020-09-27 04:50:49,088] Trial 22 finished with value: 0.6900188884483167 and parameters: {'num_leaves': 39}. Best is trial 17 with value: 0.6884769144658212.
num_leaves, val_score: 0.688477: 80%|######## | 16/20 [00:21<00:04, 1.22s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000823 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.658652 valid's binary_logloss: 0.689977
Early stopping, best iteration is:
[49] train's binary_logloss: 0.672748 valid's binary_logloss: 0.689541
num_leaves, val_score: 0.688477: 85%|########5 | 17/20 [00:22<00:03, 1.09s/it][I 2020-09-27 04:50:49,865] Trial 23 finished with value: 0.6895412052804837 and parameters: {'num_leaves': 42}. Best is trial 17 with value: 0.6884769144658212.
num_leaves, val_score: 0.688477: 85%|########5 | 17/20 [00:22<00:03, 1.09s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000785 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688532 valid's binary_logloss: 0.690109
[200] train's binary_logloss: 0.686793 valid's binary_logloss: 0.689219
[300] train's binary_logloss: 0.685512 valid's binary_logloss: 0.68876
[400] train's binary_logloss: 0.684466 valid's binary_logloss: 0.688694
[500] train's binary_logloss: 0.68348 valid's binary_logloss: 0.6886
Early stopping, best iteration is:
[438] train's binary_logloss: 0.684084 valid's binary_logloss: 0.68855
num_leaves, val_score: 0.688477: 90%|######### | 18/20 [00:24<00:02, 1.28s/it][I 2020-09-27 04:50:51,590] Trial 24 finished with value: 0.6885499863950102 and parameters: {'num_leaves': 3}. Best is trial 17 with value: 0.6884769144658212.
num_leaves, val_score: 0.688477: 90%|######### | 18/20 [00:24<00:02, 1.28s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000830 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.633629 valid's binary_logloss: 0.690048
Early stopping, best iteration is:
[41] train's binary_logloss: 0.66326 valid's binary_logloss: 0.689286
num_leaves, val_score: 0.688477: 95%|#########5| 19/20 [00:25<00:01, 1.16s/it][I 2020-09-27 04:50:52,488] Trial 25 finished with value: 0.689285779471649 and parameters: {'num_leaves': 81}. Best is trial 17 with value: 0.6884769144658212.
num_leaves, val_score: 0.688477: 95%|#########5| 19/20 [00:25<00:01, 1.16s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001270 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.669687 valid's binary_logloss: 0.68995
Early stopping, best iteration is:
[70] train's binary_logloss: 0.674818 valid's binary_logloss: 0.689845
num_leaves, val_score: 0.688477: 100%|##########| 20/20 [00:25<00:00, 1.04s/it][I 2020-09-27 04:50:53,247] Trial 26 finished with value: 0.6898451259704299 and parameters: {'num_leaves': 26}. Best is trial 17 with value: 0.6884769144658212.
num_leaves, val_score: 0.688477: 100%|##########| 20/20 [00:25<00:00, 1.30s/it]
bagging, val_score: 0.688477: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000794 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689916 valid's binary_logloss: 0.690872
[200] train's binary_logloss: 0.688688 valid's binary_logloss: 0.689969
[300] train's binary_logloss: 0.687926 valid's binary_logloss: 0.689254
[400] train's binary_logloss: 0.687375 valid's binary_logloss: 0.688976
[500] train's binary_logloss: 0.686964 valid's binary_logloss: 0.688668
[600] train's binary_logloss: 0.686637 valid's binary_logloss: 0.688603
[700] train's binary_logloss: 0.686361 valid's binary_logloss: 0.688503
[800] train's binary_logloss: 0.686138 valid's binary_logloss: 0.688427
[900] train's binary_logloss: 0.68594 valid's binary_logloss: 0.68828
Early stopping, best iteration is:
[893] train's binary_logloss: 0.685953 valid's binary_logloss: 0.688266
bagging, val_score: 0.688266: 10%|# | 1/10 [00:03<00:29, 3.27s/it][I 2020-09-27 04:50:56,524] Trial 27 finished with value: 0.6882656720836928 and parameters: {'bagging_fraction': 0.9056687704253129, 'bagging_freq': 1}. Best is trial 27 with value: 0.6882656720836928.
bagging, val_score: 0.688266: 10%|# | 1/10 [00:03<00:29, 3.27s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000821 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689949 valid's binary_logloss: 0.690893
[200] train's binary_logloss: 0.688731 valid's binary_logloss: 0.689897
[300] train's binary_logloss: 0.687963 valid's binary_logloss: 0.689438
[400] train's binary_logloss: 0.687422 valid's binary_logloss: 0.688923
[500] train's binary_logloss: 0.687012 valid's binary_logloss: 0.688674
[600] train's binary_logloss: 0.686692 valid's binary_logloss: 0.688589
[700] train's binary_logloss: 0.686428 valid's binary_logloss: 0.688421
[800] train's binary_logloss: 0.686204 valid's binary_logloss: 0.688439
[900] train's binary_logloss: 0.686006 valid's binary_logloss: 0.688398
Early stopping, best iteration is:
[868] train's binary_logloss: 0.686066 valid's binary_logloss: 0.68835
bagging, val_score: 0.688266: 20%|## | 2/10 [00:06<00:25, 3.24s/it][I 2020-09-27 04:50:59,688] Trial 28 finished with value: 0.68834951195225 and parameters: {'bagging_fraction': 0.9341933335925237, 'bagging_freq': 1}. Best is trial 27 with value: 0.6882656720836928.
bagging, val_score: 0.688266: 20%|## | 2/10 [00:06<00:25, 3.24s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000760 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689958 valid's binary_logloss: 0.690901
[200] train's binary_logloss: 0.688753 valid's binary_logloss: 0.689921
[300] train's binary_logloss: 0.687991 valid's binary_logloss: 0.68943
[400] train's binary_logloss: 0.687457 valid's binary_logloss: 0.688954
[500] train's binary_logloss: 0.687054 valid's binary_logloss: 0.688639
[600] train's binary_logloss: 0.686734 valid's binary_logloss: 0.688575
[700] train's binary_logloss: 0.686471 valid's binary_logloss: 0.688404
[800] train's binary_logloss: 0.686247 valid's binary_logloss: 0.688399
[900] train's binary_logloss: 0.686054 valid's binary_logloss: 0.688395
[1000] train's binary_logloss: 0.685884 valid's binary_logloss: 0.688386
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.685884 valid's binary_logloss: 0.688386
bagging, val_score: 0.688266: 30%|### | 3/10 [00:09<00:22, 3.22s/it][I 2020-09-27 04:51:02,884] Trial 29 finished with value: 0.6883861307718222 and parameters: {'bagging_fraction': 0.9472695270258076, 'bagging_freq': 1}. Best is trial 27 with value: 0.6882656720836928.
bagging, val_score: 0.688266: 30%|### | 3/10 [00:09<00:22, 3.22s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000818 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689959 valid's binary_logloss: 0.690908
[200] train's binary_logloss: 0.688754 valid's binary_logloss: 0.689942
[300] train's binary_logloss: 0.68799 valid's binary_logloss: 0.689383
[400] train's binary_logloss: 0.687454 valid's binary_logloss: 0.688954
[500] train's binary_logloss: 0.687052 valid's binary_logloss: 0.688747
[600] train's binary_logloss: 0.686731 valid's binary_logloss: 0.68862
[700] train's binary_logloss: 0.68647 valid's binary_logloss: 0.688482
[800] train's binary_logloss: 0.686248 valid's binary_logloss: 0.68848
Early stopping, best iteration is:
[725] train's binary_logloss: 0.686411 valid's binary_logloss: 0.68843
bagging, val_score: 0.688266: 40%|#### | 4/10 [00:12<00:18, 3.05s/it][I 2020-09-27 04:51:05,526] Trial 30 finished with value: 0.6884302974689744 and parameters: {'bagging_fraction': 0.9478078952145158, 'bagging_freq': 1}. Best is trial 27 with value: 0.6882656720836928.
bagging, val_score: 0.688266: 40%|#### | 4/10 [00:12<00:18, 3.05s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000815 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689962 valid's binary_logloss: 0.690903
[200] train's binary_logloss: 0.688757 valid's binary_logloss: 0.690012
[300] train's binary_logloss: 0.688001 valid's binary_logloss: 0.689453
[400] train's binary_logloss: 0.687468 valid's binary_logloss: 0.688994
[500] train's binary_logloss: 0.687058 valid's binary_logloss: 0.688733
[600] train's binary_logloss: 0.686738 valid's binary_logloss: 0.688599
[700] train's binary_logloss: 0.686473 valid's binary_logloss: 0.688496
[800] train's binary_logloss: 0.686255 valid's binary_logloss: 0.688458
[900] train's binary_logloss: 0.686061 valid's binary_logloss: 0.688445
Early stopping, best iteration is:
[867] train's binary_logloss: 0.686119 valid's binary_logloss: 0.688391
bagging, val_score: 0.688266: 50%|##### | 5/10 [00:15<00:15, 3.08s/it][I 2020-09-27 04:51:08,673] Trial 31 finished with value: 0.6883910883412938 and parameters: {'bagging_fraction': 0.9500225140884224, 'bagging_freq': 1}. Best is trial 27 with value: 0.6882656720836928.
bagging, val_score: 0.688266: 50%|##### | 5/10 [00:15<00:15, 3.08s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000836 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689952 valid's binary_logloss: 0.690877
[200] train's binary_logloss: 0.688739 valid's binary_logloss: 0.689932
[300] train's binary_logloss: 0.687981 valid's binary_logloss: 0.689456
[400] train's binary_logloss: 0.687443 valid's binary_logloss: 0.688946
[500] train's binary_logloss: 0.687035 valid's binary_logloss: 0.688722
[600] train's binary_logloss: 0.686712 valid's binary_logloss: 0.688567
[700] train's binary_logloss: 0.686447 valid's binary_logloss: 0.688364
[800] train's binary_logloss: 0.686224 valid's binary_logloss: 0.688372
Early stopping, best iteration is:
[701] train's binary_logloss: 0.686445 valid's binary_logloss: 0.688354
bagging, val_score: 0.688266: 60%|###### | 6/10 [00:18<00:11, 2.94s/it][I 2020-09-27 04:51:11,277] Trial 32 finished with value: 0.6883544364816402 and parameters: {'bagging_fraction': 0.939276221019466, 'bagging_freq': 1}. Best is trial 27 with value: 0.6882656720836928.
bagging, val_score: 0.688266: 60%|###### | 6/10 [00:18<00:11, 2.94s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000816 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689945 valid's binary_logloss: 0.690873
[200] train's binary_logloss: 0.688736 valid's binary_logloss: 0.68993
[300] train's binary_logloss: 0.687975 valid's binary_logloss: 0.689441
[400] train's binary_logloss: 0.687436 valid's binary_logloss: 0.688977
[500] train's binary_logloss: 0.687029 valid's binary_logloss: 0.68869
[600] train's binary_logloss: 0.68671 valid's binary_logloss: 0.688599
[700] train's binary_logloss: 0.686444 valid's binary_logloss: 0.688457
[800] train's binary_logloss: 0.686223 valid's binary_logloss: 0.688493
Early stopping, best iteration is:
[700] train's binary_logloss: 0.686444 valid's binary_logloss: 0.688457
bagging, val_score: 0.688266: 70%|####### | 7/10 [00:20<00:08, 2.83s/it][I 2020-09-27 04:51:13,857] Trial 33 finished with value: 0.6884573747067545 and parameters: {'bagging_fraction': 0.9386192260753653, 'bagging_freq': 1}. Best is trial 27 with value: 0.6882656720836928.
bagging, val_score: 0.688266: 70%|####### | 7/10 [00:20<00:08, 2.83s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000907 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689934 valid's binary_logloss: 0.690812
[200] train's binary_logloss: 0.688717 valid's binary_logloss: 0.689892
[300] train's binary_logloss: 0.687955 valid's binary_logloss: 0.689321
[400] train's binary_logloss: 0.687414 valid's binary_logloss: 0.688937
[500] train's binary_logloss: 0.686999 valid's binary_logloss: 0.688627
[600] train's binary_logloss: 0.686678 valid's binary_logloss: 0.68863
[700] train's binary_logloss: 0.686413 valid's binary_logloss: 0.688396
[800] train's binary_logloss: 0.68619 valid's binary_logloss: 0.688375
[900] train's binary_logloss: 0.685991 valid's binary_logloss: 0.688321
Early stopping, best iteration is:
[867] train's binary_logloss: 0.686053 valid's binary_logloss: 0.688263
bagging, val_score: 0.688263: 80%|######## | 8/10 [00:23<00:05, 2.91s/it][I 2020-09-27 04:51:16,952] Trial 34 finished with value: 0.6882634163638113 and parameters: {'bagging_fraction': 0.9270871288128361, 'bagging_freq': 1}. Best is trial 34 with value: 0.6882634163638113.
bagging, val_score: 0.688263: 80%|######## | 8/10 [00:23<00:05, 2.91s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000826 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627
[200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877
[300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924
[400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859
[500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858
[600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462
[700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201
Early stopping, best iteration is:
[699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186
bagging, val_score: 0.688186: 90%|######### | 9/10 [00:26<00:02, 2.83s/it][I 2020-09-27 04:51:19,581] Trial 35 finished with value: 0.6881860644541565 and parameters: {'bagging_fraction': 0.797756062798586, 'bagging_freq': 1}. Best is trial 35 with value: 0.6881860644541565.
bagging, val_score: 0.688186: 90%|######### | 9/10 [00:26<00:02, 2.83s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000754 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689842 valid's binary_logloss: 0.690437
[200] train's binary_logloss: 0.688572 valid's binary_logloss: 0.689737
[300] train's binary_logloss: 0.687777 valid's binary_logloss: 0.68909
[400] train's binary_logloss: 0.687207 valid's binary_logloss: 0.688463
[500] train's binary_logloss: 0.686786 valid's binary_logloss: 0.688389
Early stopping, best iteration is:
[463] train's binary_logloss: 0.686927 valid's binary_logloss: 0.688299
bagging, val_score: 0.688186: 100%|##########| 10/10 [00:28<00:00, 2.52s/it][I 2020-09-27 04:51:21,392] Trial 36 finished with value: 0.6882991425357883 and parameters: {'bagging_fraction': 0.7512163390719674, 'bagging_freq': 4}. Best is trial 35 with value: 0.6881860644541565.
bagging, val_score: 0.688186: 100%|##########| 10/10 [00:28<00:00, 2.81s/it]
feature_fraction_stage2, val_score: 0.688186: 0%| | 0/3 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000836 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689843 valid's binary_logloss: 0.690616
[200] train's binary_logloss: 0.688551 valid's binary_logloss: 0.689846
[300] train's binary_logloss: 0.687748 valid's binary_logloss: 0.689146
[400] train's binary_logloss: 0.687179 valid's binary_logloss: 0.688875
[500] train's binary_logloss: 0.686761 valid's binary_logloss: 0.688605
[600] train's binary_logloss: 0.686432 valid's binary_logloss: 0.68842
[700] train's binary_logloss: 0.686158 valid's binary_logloss: 0.688213
Early stopping, best iteration is:
[699] train's binary_logloss: 0.686161 valid's binary_logloss: 0.68819
feature_fraction_stage2, val_score: 0.688186: 33%|###3 | 1/3 [00:02<00:05, 2.77s/it][I 2020-09-27 04:51:24,172] Trial 37 finished with value: 0.6881903702314308 and parameters: {'feature_fraction': 0.41600000000000004}. Best is trial 37 with value: 0.6881903702314308.
feature_fraction_stage2, val_score: 0.688186: 33%|###3 | 1/3 [00:02<00:05, 2.77s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000871 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689828 valid's binary_logloss: 0.690608
[200] train's binary_logloss: 0.688553 valid's binary_logloss: 0.689888
[300] train's binary_logloss: 0.687744 valid's binary_logloss: 0.689142
[400] train's binary_logloss: 0.687173 valid's binary_logloss: 0.688816
[500] train's binary_logloss: 0.686758 valid's binary_logloss: 0.688522
[600] train's binary_logloss: 0.686423 valid's binary_logloss: 0.688429
[700] train's binary_logloss: 0.686151 valid's binary_logloss: 0.688287
Early stopping, best iteration is:
[697] train's binary_logloss: 0.686159 valid's binary_logloss: 0.688254
feature_fraction_stage2, val_score: 0.688186: 67%|######6 | 2/3 [00:05<00:02, 2.75s/it][I 2020-09-27 04:51:26,877] Trial 38 finished with value: 0.6882536571755299 and parameters: {'feature_fraction': 0.44800000000000006}. Best is trial 37 with value: 0.6881903702314308.
feature_fraction_stage2, val_score: 0.688186: 67%|######6 | 2/3 [00:05<00:02, 2.75s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000812 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689828 valid's binary_logloss: 0.690608
[200] train's binary_logloss: 0.688553 valid's binary_logloss: 0.689888
[300] train's binary_logloss: 0.687744 valid's binary_logloss: 0.689142
[400] train's binary_logloss: 0.687173 valid's binary_logloss: 0.688816
[500] train's binary_logloss: 0.686758 valid's binary_logloss: 0.688522
[600] train's binary_logloss: 0.686423 valid's binary_logloss: 0.688429
[700] train's binary_logloss: 0.686151 valid's binary_logloss: 0.688287
Early stopping, best iteration is:
[697] train's binary_logloss: 0.686159 valid's binary_logloss: 0.688254
feature_fraction_stage2, val_score: 0.688186: 100%|##########| 3/3 [00:08<00:00, 2.74s/it][I 2020-09-27 04:51:29,591] Trial 39 finished with value: 0.6882536571755299 and parameters: {'feature_fraction': 0.48000000000000004}. Best is trial 37 with value: 0.6881903702314308.
feature_fraction_stage2, val_score: 0.688186: 100%|##########| 3/3 [00:08<00:00, 2.73s/it]
regularization_factors, val_score: 0.688186: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008765 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627
[200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877
[300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924
[400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859
[500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858
[600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462
[700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201
Early stopping, best iteration is:
[699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186
regularization_factors, val_score: 0.688186: 5%|5 | 1/20 [00:02<00:48, 2.55s/it][I 2020-09-27 04:51:32,160] Trial 40 finished with value: 0.6881860967213117 and parameters: {'lambda_l1': 4.739533108597076e-06, 'lambda_l2': 0.011058449085898892}. Best is trial 40 with value: 0.6881860967213117.
regularization_factors, val_score: 0.688186: 5%|5 | 1/20 [00:02<00:48, 2.55s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000761 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627
[200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877
[300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924
[400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859
[500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858
[600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462
[700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201
Early stopping, best iteration is:
[699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186
regularization_factors, val_score: 0.688186: 10%|# | 2/20 [00:05<00:46, 2.59s/it][I 2020-09-27 04:51:34,825] Trial 41 finished with value: 0.6881861017696784 and parameters: {'lambda_l1': 2.6160452041031395e-06, 'lambda_l2': 0.012798063629265445}. Best is trial 40 with value: 0.6881860967213117.
regularization_factors, val_score: 0.688186: 10%|# | 2/20 [00:05<00:46, 2.59s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000802 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627
[200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877
[300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924
[400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859
[500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858
[600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462
[700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201
Early stopping, best iteration is:
[699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186
regularization_factors, val_score: 0.688186: 15%|#5 | 3/20 [00:07<00:44, 2.59s/it][I 2020-09-27 04:51:37,430] Trial 42 finished with value: 0.6881860972873897 and parameters: {'lambda_l1': 2.1069017715929826e-06, 'lambda_l2': 0.011265308763779644}. Best is trial 40 with value: 0.6881860967213117.
regularization_factors, val_score: 0.688186: 15%|#5 | 3/20 [00:07<00:44, 2.59s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000812 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627
[200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877
[300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924
[400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859
[500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858
[600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462
[700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201
Early stopping, best iteration is:
[699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186
regularization_factors, val_score: 0.688186: 20%|## | 4/20 [00:10<00:42, 2.64s/it][I 2020-09-27 04:51:40,183] Trial 43 finished with value: 0.6881861188350943 and parameters: {'lambda_l1': 2.871077466536222e-06, 'lambda_l2': 0.01863739043244412}. Best is trial 40 with value: 0.6881860967213117.
regularization_factors, val_score: 0.688186: 20%|## | 4/20 [00:10<00:42, 2.64s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000761 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627
[200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877
[300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924
[400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859
[500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858
[600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462
[700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201
Early stopping, best iteration is:
[699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186
regularization_factors, val_score: 0.688186: 25%|##5 | 5/20 [00:13<00:39, 2.65s/it][I 2020-09-27 04:51:42,840] Trial 44 finished with value: 0.6881860987513834 and parameters: {'lambda_l1': 1.3228363853250858e-06, 'lambda_l2': 0.011769235325842804}. Best is trial 40 with value: 0.6881860967213117.
regularization_factors, val_score: 0.688186: 25%|##5 | 5/20 [00:13<00:39, 2.65s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000756 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627
[200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877
[300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924
[400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859
[500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858
[600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462
[700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201
Early stopping, best iteration is:
[699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186
regularization_factors, val_score: 0.688186: 30%|### | 6/20 [00:15<00:37, 2.66s/it][I 2020-09-27 04:51:45,544] Trial 45 finished with value: 0.6881861035517631 and parameters: {'lambda_l1': 1.083504577056653e-06, 'lambda_l2': 0.013413611446410094}. Best is trial 40 with value: 0.6881860967213117.
regularization_factors, val_score: 0.688186: 30%|### | 6/20 [00:15<00:37, 2.66s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000815 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627
[200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877
[300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924
[400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859
[500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858
[600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462
[700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201
Early stopping, best iteration is:
[699] train's binary_logloss: 0.686174 valid's binary_logloss: 0.688186
regularization_factors, val_score: 0.688186: 35%|###5 | 7/20 [00:18<00:34, 2.64s/it][I 2020-09-27 04:51:48,142] Trial 46 finished with value: 0.688186150142476 and parameters: {'lambda_l1': 3.3314910180930466e-06, 'lambda_l2': 0.029327749830602143}. Best is trial 40 with value: 0.6881860967213117.
regularization_factors, val_score: 0.688186: 35%|###5 | 7/20 [00:18<00:34, 2.64s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000818 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627
[200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877
[300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924
[400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859
[500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858
[600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462
[700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201
Early stopping, best iteration is:
[699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186
regularization_factors, val_score: 0.688186: 40%|#### | 8/20 [00:21<00:31, 2.65s/it][I 2020-09-27 04:51:50,799] Trial 47 finished with value: 0.6881860757482499 and parameters: {'lambda_l1': 2.3526621184608058e-06, 'lambda_l2': 0.003868598725242898}. Best is trial 47 with value: 0.6881860757482499.
regularization_factors, val_score: 0.688186: 40%|#### | 8/20 [00:21<00:31, 2.65s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000808 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627
[200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877
[300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924
[400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859
[500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858
[600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462
[700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201
Early stopping, best iteration is:
[699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186
regularization_factors, val_score: 0.688186: 45%|####5 | 9/20 [00:23<00:29, 2.65s/it][I 2020-09-27 04:51:53,467] Trial 48 finished with value: 0.6881860645612168 and parameters: {'lambda_l1': 3.130361329176902e-06, 'lambda_l2': 2.312087319678873e-05}. Best is trial 48 with value: 0.6881860645612168.
regularization_factors, val_score: 0.688186: 45%|####5 | 9/20 [00:23<00:29, 2.65s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000816 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627
[200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877
[300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924
[400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859
[500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858
[600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462
[700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201
Early stopping, best iteration is:
[699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186
regularization_factors, val_score: 0.688186: 50%|##### | 10/20 [00:26<00:26, 2.65s/it][I 2020-09-27 04:51:56,110] Trial 49 finished with value: 0.6881860867937425 and parameters: {'lambda_l1': 0.0016839304778010226, 'lambda_l2': 1.0716481604934569e-06}. Best is trial 48 with value: 0.6881860645612168.
regularization_factors, val_score: 0.688186: 50%|##### | 10/20 [00:26<00:26, 2.65s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000807 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689854 valid's binary_logloss: 0.690628
[200] train's binary_logloss: 0.688573 valid's binary_logloss: 0.689878
[300] train's binary_logloss: 0.687778 valid's binary_logloss: 0.689241
[400] train's binary_logloss: 0.687211 valid's binary_logloss: 0.688892
[500] train's binary_logloss: 0.686781 valid's binary_logloss: 0.688591
[600] train's binary_logloss: 0.686449 valid's binary_logloss: 0.688474
[700] train's binary_logloss: 0.686174 valid's binary_logloss: 0.688261
[800] train's binary_logloss: 0.685939 valid's binary_logloss: 0.688313
Early stopping, best iteration is:
[739] train's binary_logloss: 0.686081 valid's binary_logloss: 0.688233
regularization_factors, val_score: 0.688186: 55%|#####5 | 11/20 [00:29<00:24, 2.69s/it][I 2020-09-27 04:51:58,894] Trial 50 finished with value: 0.6882329753342665 and parameters: {'lambda_l1': 0.09692234973226965, 'lambda_l2': 8.775766596471004e-07}. Best is trial 48 with value: 0.6881860645612168.
regularization_factors, val_score: 0.688186: 55%|#####5 | 11/20 [00:29<00:24, 2.69s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000815 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627
[200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877
[300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924
[400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859
[500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858
[600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462
[700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201
Early stopping, best iteration is:
[699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186
regularization_factors, val_score: 0.688186: 60%|###### | 12/20 [00:31<00:21, 2.68s/it][I 2020-09-27 04:52:01,543] Trial 51 finished with value: 0.6881861075194643 and parameters: {'lambda_l1': 0.0032449073251714445, 'lambda_l2': 3.2716223509922217e-06}. Best is trial 48 with value: 0.6881860645612168.
regularization_factors, val_score: 0.688186: 60%|###### | 12/20 [00:31<00:21, 2.68s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000769 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627
[200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877
[300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924
[400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859
[500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858
[600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462
[700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201
Early stopping, best iteration is:
[699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186
regularization_factors, val_score: 0.688186: 65%|######5 | 13/20 [00:34<00:18, 2.67s/it][I 2020-09-27 04:52:04,186] Trial 52 finished with value: 0.6881860698451091 and parameters: {'lambda_l1': 0.0003848047657014028, 'lambda_l2': 9.591675343195672e-05}. Best is trial 48 with value: 0.6881860645612168.
regularization_factors, val_score: 0.688186: 65%|######5 | 13/20 [00:34<00:18, 2.67s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000847 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627
[200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877
[300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924
[400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859
[500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858
[600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462
[700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201
Early stopping, best iteration is:
[699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186
regularization_factors, val_score: 0.688186: 70%|####### | 14/20 [00:37<00:15, 2.65s/it][I 2020-09-27 04:52:06,789] Trial 53 finished with value: 0.6881860703621502 and parameters: {'lambda_l1': 0.0004165194013850262, 'lambda_l2': 0.0001302229603956366}. Best is trial 48 with value: 0.6881860645612168.
regularization_factors, val_score: 0.688186: 70%|####### | 14/20 [00:37<00:15, 2.65s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000813 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627
[200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877
[300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924
[400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859
[500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858
[600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462
[700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201
Early stopping, best iteration is:
[699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186
regularization_factors, val_score: 0.688186: 75%|#######5 | 15/20 [00:42<00:17, 3.49s/it][I 2020-09-27 04:52:12,308] Trial 54 finished with value: 0.6881860706975759 and parameters: {'lambda_l1': 0.00045861316138919045, 'lambda_l2': 5.257878461523017e-05}. Best is trial 48 with value: 0.6881860645612168.
regularization_factors, val_score: 0.688186: 75%|#######5 | 15/20 [00:42<00:17, 3.49s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001772 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627
[200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877
[300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924
[400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859
[500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858
[600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462
[700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201
Early stopping, best iteration is:
[699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186
regularization_factors, val_score: 0.688186: 80%|######## | 16/20 [00:47<00:15, 3.79s/it][I 2020-09-27 04:52:16,741] Trial 55 finished with value: 0.6881860672245419 and parameters: {'lambda_l1': 0.00019089434047484262, 'lambda_l2': 8.205565377180615e-05}. Best is trial 48 with value: 0.6881860645612168.
regularization_factors, val_score: 0.688186: 80%|######## | 16/20 [00:47<00:15, 3.79s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000880 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627
[200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877
[300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924
[400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859
[500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858
[600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462
[700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201
Early stopping, best iteration is:
[699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186
regularization_factors, val_score: 0.688186: 85%|########5 | 17/20 [00:50<00:10, 3.55s/it][I 2020-09-27 04:52:19,720] Trial 56 finished with value: 0.6881860669984095 and parameters: {'lambda_l1': 0.00017461333741205492, 'lambda_l2': 7.853596223254834e-05}. Best is trial 48 with value: 0.6881860645612168.
regularization_factors, val_score: 0.688186: 85%|########5 | 17/20 [00:50<00:10, 3.55s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008942 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627
[200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877
[300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924
[400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859
[500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858
[600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462
[700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201
Early stopping, best iteration is:
[699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186
regularization_factors, val_score: 0.688186: 90%|######### | 18/20 [00:53<00:06, 3.38s/it][I 2020-09-27 04:52:22,709] Trial 57 finished with value: 0.6881860668888051 and parameters: {'lambda_l1': 0.0001634804643071105, 'lambda_l2': 9.030958952909047e-05}. Best is trial 48 with value: 0.6881860645612168.
regularization_factors, val_score: 0.688186: 90%|######### | 18/20 [00:53<00:06, 3.38s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000820 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627
[200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877
[300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924
[400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859
[500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858
[600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462
[700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201
Early stopping, best iteration is:
[699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186
regularization_factors, val_score: 0.688186: 95%|#########5| 19/20 [00:56<00:03, 3.29s/it][I 2020-09-27 04:52:25,792] Trial 58 finished with value: 0.6881860653671973 and parameters: {'lambda_l1': 5.5880426121768656e-05, 'lambda_l2': 5.7285984514290246e-05}. Best is trial 48 with value: 0.6881860645612168.
regularization_factors, val_score: 0.688186: 95%|#########5| 19/20 [00:56<00:03, 3.29s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001313 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627
[200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877
[300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924
[400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859
[500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858
[600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462
[700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201
Early stopping, best iteration is:
[699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186
regularization_factors, val_score: 0.688186: 100%|##########| 20/20 [00:59<00:00, 3.27s/it][I 2020-09-27 04:52:28,997] Trial 59 finished with value: 0.6881860649508688 and parameters: {'lambda_l1': 3.502289401416132e-05, 'lambda_l2': 1.1183876599807051e-05}. Best is trial 48 with value: 0.6881860645612168.
regularization_factors, val_score: 0.688186: 100%|##########| 20/20 [00:59<00:00, 2.97s/it]
min_data_in_leaf, val_score: 0.688186: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000910 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627
[200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877
[300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924
[400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859
[500] train's binary_logloss: 0.686783 valid's binary_logloss: 0.688582
[600] train's binary_logloss: 0.686449 valid's binary_logloss: 0.688469
[700] train's binary_logloss: 0.686176 valid's binary_logloss: 0.688262
Early stopping, best iteration is:
[699] train's binary_logloss: 0.686178 valid's binary_logloss: 0.688248
min_data_in_leaf, val_score: 0.688186: 20%|## | 1/5 [00:02<00:11, 2.91s/it][I 2020-09-27 04:52:31,924] Trial 60 finished with value: 0.6882475377052306 and parameters: {'min_child_samples': 25}. Best is trial 60 with value: 0.6882475377052306.
min_data_in_leaf, val_score: 0.688186: 20%|## | 1/5 [00:02<00:11, 2.91s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000804 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627
[200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877
[300] train's binary_logloss: 0.687769 valid's binary_logloss: 0.689167
[400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688804
[500] train's binary_logloss: 0.68679 valid's binary_logloss: 0.688551
[600] train's binary_logloss: 0.686455 valid's binary_logloss: 0.688496
[700] train's binary_logloss: 0.686182 valid's binary_logloss: 0.68822
Early stopping, best iteration is:
[699] train's binary_logloss: 0.686183 valid's binary_logloss: 0.688205
min_data_in_leaf, val_score: 0.688186: 40%|#### | 2/5 [00:05<00:08, 2.90s/it][I 2020-09-27 04:52:34,798] Trial 61 finished with value: 0.688205087532443 and parameters: {'min_child_samples': 50}. Best is trial 61 with value: 0.688205087532443.
min_data_in_leaf, val_score: 0.688186: 40%|#### | 2/5 [00:05<00:08, 2.90s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008806 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627
[200] train's binary_logloss: 0.688564 valid's binary_logloss: 0.689845
[300] train's binary_logloss: 0.687763 valid's binary_logloss: 0.689203
[400] train's binary_logloss: 0.687188 valid's binary_logloss: 0.688823
[500] train's binary_logloss: 0.686762 valid's binary_logloss: 0.688533
[600] train's binary_logloss: 0.686429 valid's binary_logloss: 0.688447
[700] train's binary_logloss: 0.686152 valid's binary_logloss: 0.688247
Early stopping, best iteration is:
[699] train's binary_logloss: 0.686153 valid's binary_logloss: 0.688233
min_data_in_leaf, val_score: 0.688186: 60%|###### | 3/5 [00:08<00:05, 2.89s/it][I 2020-09-27 04:52:37,667] Trial 62 finished with value: 0.688232682196116 and parameters: {'min_child_samples': 10}. Best is trial 61 with value: 0.688205087532443.
min_data_in_leaf, val_score: 0.688186: 60%|###### | 3/5 [00:08<00:05, 2.89s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000818 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627
[200] train's binary_logloss: 0.688564 valid's binary_logloss: 0.689845
[300] train's binary_logloss: 0.687763 valid's binary_logloss: 0.689203
[400] train's binary_logloss: 0.687188 valid's binary_logloss: 0.688823
[500] train's binary_logloss: 0.686762 valid's binary_logloss: 0.688533
[600] train's binary_logloss: 0.686429 valid's binary_logloss: 0.688447
[700] train's binary_logloss: 0.686152 valid's binary_logloss: 0.688247
[800] train's binary_logloss: 0.68591 valid's binary_logloss: 0.688233
[900] train's binary_logloss: 0.685702 valid's binary_logloss: 0.688215
[1000] train's binary_logloss: 0.685499 valid's binary_logloss: 0.688212
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.685499 valid's binary_logloss: 0.688212
min_data_in_leaf, val_score: 0.688186: 80%|######## | 4/5 [00:12<00:03, 3.15s/it][I 2020-09-27 04:52:41,438] Trial 63 finished with value: 0.6882121446078117 and parameters: {'min_child_samples': 5}. Best is trial 61 with value: 0.688205087532443.
min_data_in_leaf, val_score: 0.688186: 80%|######## | 4/5 [00:12<00:03, 3.15s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000805 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040
[LightGBM] [Info] Start training from score 0.010040
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689902 valid's binary_logloss: 0.690519
[200] train's binary_logloss: 0.68864 valid's binary_logloss: 0.68968
[300] train's binary_logloss: 0.687867 valid's binary_logloss: 0.688977
[400] train's binary_logloss: 0.687339 valid's binary_logloss: 0.688725
[500] train's binary_logloss: 0.686927 valid's binary_logloss: 0.688484
[600] train's binary_logloss: 0.686611 valid's binary_logloss: 0.688434
Early stopping, best iteration is:
[531] train's binary_logloss: 0.68682 valid's binary_logloss: 0.688362
min_data_in_leaf, val_score: 0.688186: 100%|##########| 5/5 [00:15<00:00, 3.03s/it][I 2020-09-27 04:52:44,170] Trial 64 finished with value: 0.6883621338709048 and parameters: {'min_child_samples': 100}. Best is trial 61 with value: 0.688205087532443.
min_data_in_leaf, val_score: 0.688186: 100%|##########| 5/5 [00:15<00:00, 3.03s/it]
Fold : 2
[I 2020-09-27 04:52:44,427] A new study created in memory with name: no-name-4beb476a-e1cd-441a-9d78-1d4a1c5ec6fe
feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010680 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663809 valid's binary_logloss: 0.689691
Early stopping, best iteration is:
[55] train's binary_logloss: 0.674139 valid's binary_logloss: 0.689343
feature_fraction, val_score: 0.689343: 14%|#4 | 1/7 [00:01<00:09, 1.55s/it][I 2020-09-27 04:52:45,997] Trial 0 finished with value: 0.689343019442042 and parameters: {'feature_fraction': 0.7}. Best is trial 0 with value: 0.689343019442042.
feature_fraction, val_score: 0.689343: 14%|#4 | 1/7 [00:01<00:09, 1.55s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008342 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664318 valid's binary_logloss: 0.688991
Early stopping, best iteration is:
[63] train's binary_logloss: 0.672462 valid's binary_logloss: 0.688554
feature_fraction, val_score: 0.688554: 29%|##8 | 2/7 [00:03<00:07, 1.58s/it][I 2020-09-27 04:52:47,648] Trial 1 finished with value: 0.6885538121639901 and parameters: {'feature_fraction': 0.6}. Best is trial 1 with value: 0.6885538121639901.
feature_fraction, val_score: 0.688554: 29%|##8 | 2/7 [00:03<00:07, 1.58s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001457 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.666146 valid's binary_logloss: 0.68872
Early stopping, best iteration is:
[99] train's binary_logloss: 0.66631 valid's binary_logloss: 0.688654
feature_fraction, val_score: 0.688554: 43%|####2 | 3/7 [00:05<00:06, 1.68s/it][I 2020-09-27 04:52:49,556] Trial 2 finished with value: 0.6886536923224554 and parameters: {'feature_fraction': 0.4}. Best is trial 1 with value: 0.6885538121639901.
feature_fraction, val_score: 0.688554: 43%|####2 | 3/7 [00:05<00:06, 1.68s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001582 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811
[200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036
Early stopping, best iteration is:
[128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306
feature_fraction, val_score: 0.688306: 57%|#####7 | 4/7 [00:06<00:04, 1.56s/it][I 2020-09-27 04:52:50,831] Trial 3 finished with value: 0.6883061543757897 and parameters: {'feature_fraction': 0.5}. Best is trial 3 with value: 0.6883061543757897.
feature_fraction, val_score: 0.688306: 57%|#####7 | 4/7 [00:06<00:04, 1.56s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013980 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663334 valid's binary_logloss: 0.689928
Early stopping, best iteration is:
[39] train's binary_logloss: 0.678023 valid's binary_logloss: 0.689633
feature_fraction, val_score: 0.688306: 71%|#######1 | 5/7 [00:07<00:02, 1.33s/it][I 2020-09-27 04:52:51,651] Trial 4 finished with value: 0.6896333450899451 and parameters: {'feature_fraction': 0.8}. Best is trial 3 with value: 0.6883061543757897.
feature_fraction, val_score: 0.688306: 71%|#######1 | 5/7 [00:07<00:02, 1.33s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.003355 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.662709 valid's binary_logloss: 0.690225
Early stopping, best iteration is:
[72] train's binary_logloss: 0.669305 valid's binary_logloss: 0.689574
feature_fraction, val_score: 0.688306: 86%|########5 | 6/7 [00:08<00:01, 1.26s/it][I 2020-09-27 04:52:52,738] Trial 5 finished with value: 0.6895740757388882 and parameters: {'feature_fraction': 1.0}. Best is trial 3 with value: 0.6883061543757897.
feature_fraction, val_score: 0.688306: 86%|########5 | 6/7 [00:08<00:01, 1.26s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001567 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663043 valid's binary_logloss: 0.689123
Early stopping, best iteration is:
[74] train's binary_logloss: 0.668993 valid's binary_logloss: 0.688769
feature_fraction, val_score: 0.688306: 100%|##########| 7/7 [00:09<00:00, 1.20s/it][I 2020-09-27 04:52:53,795] Trial 6 finished with value: 0.6887687471521096 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 3 with value: 0.6883061543757897.
feature_fraction, val_score: 0.688306: 100%|##########| 7/7 [00:09<00:00, 1.34s/it]
num_leaves, val_score: 0.688306: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010710 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.642699 valid's binary_logloss: 0.689465
Early stopping, best iteration is:
[60] train's binary_logloss: 0.658614 valid's binary_logloss: 0.688722
num_leaves, val_score: 0.688306: 5%|5 | 1/20 [00:01<00:19, 1.05s/it][I 2020-09-27 04:52:54,861] Trial 7 finished with value: 0.6887216519277777 and parameters: {'num_leaves': 63}. Best is trial 7 with value: 0.6887216519277777.
num_leaves, val_score: 0.688306: 5%|5 | 1/20 [00:01<00:19, 1.05s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001033 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.541922 valid's binary_logloss: 0.696672
Early stopping, best iteration is:
[17] train's binary_logloss: 0.656043 valid's binary_logloss: 0.691091
num_leaves, val_score: 0.688306: 10%|# | 2/20 [00:02<00:21, 1.17s/it][I 2020-09-27 04:52:56,310] Trial 8 finished with value: 0.6910909448392947 and parameters: {'num_leaves': 247}. Best is trial 7 with value: 0.6887216519277777.
num_leaves, val_score: 0.688306: 10%|# | 2/20 [00:02<00:21, 1.17s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000948 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.661372 valid's binary_logloss: 0.689877
Early stopping, best iteration is:
[75] train's binary_logloss: 0.667459 valid's binary_logloss: 0.689409
num_leaves, val_score: 0.688306: 15%|#5 | 3/20 [00:03<00:18, 1.09s/it][I 2020-09-27 04:52:57,203] Trial 9 finished with value: 0.6894087377385544 and parameters: {'num_leaves': 36}. Best is trial 7 with value: 0.6887216519277777.
num_leaves, val_score: 0.688306: 15%|#5 | 3/20 [00:03<00:18, 1.09s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000877 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.543852 valid's binary_logloss: 0.693135
Early stopping, best iteration is:
[20] train's binary_logloss: 0.651184 valid's binary_logloss: 0.690432
num_leaves, val_score: 0.688306: 20%|## | 4/20 [00:04<00:18, 1.17s/it][I 2020-09-27 04:52:58,582] Trial 10 finished with value: 0.6904319926657629 and parameters: {'num_leaves': 243}. Best is trial 7 with value: 0.6887216519277777.
num_leaves, val_score: 0.688306: 20%|## | 4/20 [00:04<00:18, 1.17s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000959 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.589007 valid's binary_logloss: 0.693108
Early stopping, best iteration is:
[37] train's binary_logloss: 0.644233 valid's binary_logloss: 0.690224
num_leaves, val_score: 0.688306: 25%|##5 | 5/20 [00:05<00:17, 1.19s/it][I 2020-09-27 04:52:59,794] Trial 11 finished with value: 0.6902240196479881 and parameters: {'num_leaves': 155}. Best is trial 7 with value: 0.6887216519277777.
num_leaves, val_score: 0.688306: 25%|##5 | 5/20 [00:05<00:17, 1.19s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000952 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.596218 valid's binary_logloss: 0.693075
Early stopping, best iteration is:
[34] train's binary_logloss: 0.650962 valid's binary_logloss: 0.689486
num_leaves, val_score: 0.688306: 30%|### | 6/20 [00:07<00:16, 1.17s/it][I 2020-09-27 04:53:00,935] Trial 12 finished with value: 0.689485631745598 and parameters: {'num_leaves': 141}. Best is trial 7 with value: 0.6887216519277777.
num_leaves, val_score: 0.688306: 30%|### | 6/20 [00:07<00:16, 1.17s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000944 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.69009 valid's binary_logloss: 0.690453
[200] train's binary_logloss: 0.688904 valid's binary_logloss: 0.689731
[300] train's binary_logloss: 0.688162 valid's binary_logloss: 0.689318
[400] train's binary_logloss: 0.687642 valid's binary_logloss: 0.688959
[500] train's binary_logloss: 0.687262 valid's binary_logloss: 0.688713
[600] train's binary_logloss: 0.686964 valid's binary_logloss: 0.688609
[700] train's binary_logloss: 0.686721 valid's binary_logloss: 0.688505
[800] train's binary_logloss: 0.686514 valid's binary_logloss: 0.688439
[900] train's binary_logloss: 0.686339 valid's binary_logloss: 0.688402
[1000] train's binary_logloss: 0.686188 valid's binary_logloss: 0.688357
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.686188 valid's binary_logloss: 0.688357
num_leaves, val_score: 0.688306: 35%|###5 | 7/20 [00:10<00:23, 1.78s/it][I 2020-09-27 04:53:04,147] Trial 13 finished with value: 0.6883565901301231 and parameters: {'num_leaves': 2}. Best is trial 13 with value: 0.6883565901301231.
num_leaves, val_score: 0.688306: 35%|###5 | 7/20 [00:10<00:23, 1.78s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001010 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.67672 valid's binary_logloss: 0.689306
[200] train's binary_logloss: 0.666267 valid's binary_logloss: 0.689375
Early stopping, best iteration is:
[107] train's binary_logloss: 0.675908 valid's binary_logloss: 0.689164
num_leaves, val_score: 0.688306: 40%|#### | 8/20 [00:11<00:18, 1.53s/it][I 2020-09-27 04:53:05,090] Trial 14 finished with value: 0.6891638737730074 and parameters: {'num_leaves': 16}. Best is trial 13 with value: 0.6883565901301231.
num_leaves, val_score: 0.688306: 40%|#### | 8/20 [00:11<00:18, 1.53s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000937 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.625772 valid's binary_logloss: 0.691507
Early stopping, best iteration is:
[69] train's binary_logloss: 0.642008 valid's binary_logloss: 0.6905
num_leaves, val_score: 0.688306: 45%|####5 | 9/20 [00:12<00:15, 1.42s/it][I 2020-09-27 04:53:06,262] Trial 15 finished with value: 0.6904995062634406 and parameters: {'num_leaves': 90}. Best is trial 13 with value: 0.6883565901301231.
num_leaves, val_score: 0.688306: 45%|####5 | 9/20 [00:12<00:15, 1.42s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001295 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.577877 valid's binary_logloss: 0.694164
Early stopping, best iteration is:
[28] train's binary_logloss: 0.65064 valid's binary_logloss: 0.689958
num_leaves, val_score: 0.688306: 50%|##### | 10/20 [00:13<00:13, 1.38s/it][I 2020-09-27 04:53:07,528] Trial 16 finished with value: 0.6899575658354835 and parameters: {'num_leaves': 175}. Best is trial 13 with value: 0.6883565901301231.
num_leaves, val_score: 0.688306: 50%|##### | 10/20 [00:13<00:13, 1.38s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000927 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.620139 valid's binary_logloss: 0.691195
Early stopping, best iteration is:
[42] train's binary_logloss: 0.655396 valid's binary_logloss: 0.689705
num_leaves, val_score: 0.688306: 55%|#####5 | 11/20 [00:14<00:11, 1.26s/it][I 2020-09-27 04:53:08,520] Trial 17 finished with value: 0.6897051455072504 and parameters: {'num_leaves': 99}. Best is trial 13 with value: 0.6883565901301231.
num_leaves, val_score: 0.688306: 55%|#####5 | 11/20 [00:14<00:11, 1.26s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007712 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.686259 valid's binary_logloss: 0.689494
[200] train's binary_logloss: 0.683128 valid's binary_logloss: 0.689035
[300] train's binary_logloss: 0.680538 valid's binary_logloss: 0.68913
Early stopping, best iteration is:
[237] train's binary_logloss: 0.682127 valid's binary_logloss: 0.688927
num_leaves, val_score: 0.688306: 60%|###### | 12/20 [00:15<00:09, 1.22s/it][I 2020-09-27 04:53:09,653] Trial 18 finished with value: 0.6889270132361176 and parameters: {'num_leaves': 5}. Best is trial 13 with value: 0.6883565901301231.
num_leaves, val_score: 0.688306: 60%|###### | 12/20 [00:15<00:09, 1.22s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001555 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.566389 valid's binary_logloss: 0.69472
Early stopping, best iteration is:
[21] train's binary_logloss: 0.655891 valid's binary_logloss: 0.689891
num_leaves, val_score: 0.688306: 65%|######5 | 13/20 [00:17<00:08, 1.23s/it][I 2020-09-27 04:53:10,888] Trial 19 finished with value: 0.6898906845392911 and parameters: {'num_leaves': 199}. Best is trial 13 with value: 0.6883565901301231.
num_leaves, val_score: 0.688306: 65%|######5 | 13/20 [00:17<00:08, 1.23s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000971 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.649368 valid's binary_logloss: 0.68967
Early stopping, best iteration is:
[86] train's binary_logloss: 0.65419 valid's binary_logloss: 0.689079
num_leaves, val_score: 0.688306: 70%|####### | 14/20 [00:18<00:07, 1.17s/it][I 2020-09-27 04:53:11,922] Trial 20 finished with value: 0.689079067247648 and parameters: {'num_leaves': 53}. Best is trial 13 with value: 0.6883565901301231.
num_leaves, val_score: 0.688306: 70%|####### | 14/20 [00:18<00:07, 1.17s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001094 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.616545 valid's binary_logloss: 0.691175
Early stopping, best iteration is:
[24] train's binary_logloss: 0.668053 valid's binary_logloss: 0.690395
num_leaves, val_score: 0.688306: 75%|#######5 | 15/20 [00:19<00:05, 1.10s/it][I 2020-09-27 04:53:12,870] Trial 21 finished with value: 0.6903952269669797 and parameters: {'num_leaves': 105}. Best is trial 13 with value: 0.6883565901301231.
num_leaves, val_score: 0.688306: 75%|#######5 | 15/20 [00:19<00:05, 1.10s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000957 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.560451 valid's binary_logloss: 0.695619
Early stopping, best iteration is:
[21] train's binary_logloss: 0.654473 valid's binary_logloss: 0.690005
num_leaves, val_score: 0.688306: 80%|######## | 16/20 [00:20<00:04, 1.16s/it][I 2020-09-27 04:53:14,161] Trial 22 finished with value: 0.6900048045676193 and parameters: {'num_leaves': 209}. Best is trial 13 with value: 0.6883565901301231.
num_leaves, val_score: 0.688306: 80%|######## | 16/20 [00:20<00:04, 1.16s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000964 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.683466 valid's binary_logloss: 0.688908
[200] train's binary_logloss: 0.678323 valid's binary_logloss: 0.688588
Early stopping, best iteration is:
[128] train's binary_logloss: 0.681889 valid's binary_logloss: 0.688484
num_leaves, val_score: 0.688306: 85%|########5 | 17/20 [00:21<00:03, 1.08s/it][I 2020-09-27 04:53:15,053] Trial 23 finished with value: 0.68848442653872 and parameters: {'num_leaves': 8}. Best is trial 13 with value: 0.6883565901301231.
num_leaves, val_score: 0.688306: 85%|########5 | 17/20 [00:21<00:03, 1.08s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000948 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.677505 valid's binary_logloss: 0.68888
[200] train's binary_logloss: 0.667867 valid's binary_logloss: 0.689195
Early stopping, best iteration is:
[101] train's binary_logloss: 0.677397 valid's binary_logloss: 0.688863
num_leaves, val_score: 0.688306: 90%|######### | 18/20 [00:22<00:02, 1.01s/it][I 2020-09-27 04:53:15,914] Trial 24 finished with value: 0.6888633696872527 and parameters: {'num_leaves': 15}. Best is trial 13 with value: 0.6883565901301231.
num_leaves, val_score: 0.688306: 90%|######### | 18/20 [00:22<00:02, 1.01s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000957 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687323 valid's binary_logloss: 0.689724
[200] train's binary_logloss: 0.684864 valid's binary_logloss: 0.68927
[300] train's binary_logloss: 0.682878 valid's binary_logloss: 0.688942
[400] train's binary_logloss: 0.681166 valid's binary_logloss: 0.688902
Early stopping, best iteration is:
[348] train's binary_logloss: 0.682019 valid's binary_logloss: 0.688739
num_leaves, val_score: 0.688306: 95%|#########5| 19/20 [00:23<00:01, 1.18s/it][I 2020-09-27 04:53:17,497] Trial 25 finished with value: 0.688738728369025 and parameters: {'num_leaves': 4}. Best is trial 13 with value: 0.6883565901301231.
num_leaves, val_score: 0.688306: 95%|#########5| 19/20 [00:23<00:01, 1.18s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000982 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.638699 valid's binary_logloss: 0.691078
Early stopping, best iteration is:
[34] train's binary_logloss: 0.669389 valid's binary_logloss: 0.689921
num_leaves, val_score: 0.688306: 100%|##########| 20/20 [00:24<00:00, 1.09s/it][I 2020-09-27 04:53:18,377] Trial 26 finished with value: 0.6899211705265006 and parameters: {'num_leaves': 69}. Best is trial 13 with value: 0.6883565901301231.
num_leaves, val_score: 0.688306: 100%|##########| 20/20 [00:24<00:00, 1.23s/it]
bagging, val_score: 0.688306: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001241 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.66507 valid's binary_logloss: 0.68924
Early stopping, best iteration is:
[81] train's binary_logloss: 0.669105 valid's binary_logloss: 0.689037
bagging, val_score: 0.688306: 10%|# | 1/10 [00:01<00:09, 1.05s/it][I 2020-09-27 04:53:19,441] Trial 27 finished with value: 0.6890373745262042 and parameters: {'bagging_fraction': 0.918366805969969, 'bagging_freq': 7}. Best is trial 27 with value: 0.6890373745262042.
bagging, val_score: 0.688306: 10%|# | 1/10 [00:01<00:09, 1.05s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009692 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.666459 valid's binary_logloss: 0.689984
Early stopping, best iteration is:
[75] train's binary_logloss: 0.671612 valid's binary_logloss: 0.68934
bagging, val_score: 0.688306: 20%|## | 2/10 [00:02<00:08, 1.03s/it][I 2020-09-27 04:53:20,408] Trial 28 finished with value: 0.6893396192520514 and parameters: {'bagging_fraction': 0.4219518026182199, 'bagging_freq': 1}. Best is trial 27 with value: 0.6890373745262042.
bagging, val_score: 0.688306: 20%|## | 2/10 [00:02<00:08, 1.03s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000951 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.666862 valid's binary_logloss: 0.691464
Early stopping, best iteration is:
[53] train's binary_logloss: 0.676595 valid's binary_logloss: 0.69025
bagging, val_score: 0.688306: 30%|### | 3/10 [00:02<00:06, 1.04it/s][I 2020-09-27 04:53:21,234] Trial 29 finished with value: 0.6902501595666304 and parameters: {'bagging_fraction': 0.4030288821219972, 'bagging_freq': 4}. Best is trial 27 with value: 0.6890373745262042.
bagging, val_score: 0.688306: 30%|### | 3/10 [00:02<00:06, 1.04it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000959 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665081 valid's binary_logloss: 0.689749
[200] train's binary_logloss: 0.645221 valid's binary_logloss: 0.689436
Early stopping, best iteration is:
[181] train's binary_logloss: 0.648839 valid's binary_logloss: 0.689363
bagging, val_score: 0.688306: 40%|#### | 4/10 [00:04<00:06, 1.10s/it][I 2020-09-27 04:53:22,634] Trial 30 finished with value: 0.6893631828354185 and parameters: {'bagging_fraction': 0.7296879278779264, 'bagging_freq': 1}. Best is trial 27 with value: 0.6890373745262042.
bagging, val_score: 0.688306: 40%|#### | 4/10 [00:04<00:06, 1.10s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000952 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664945 valid's binary_logloss: 0.689235
Early stopping, best iteration is:
[56] train's binary_logloss: 0.674807 valid's binary_logloss: 0.688819
bagging, val_score: 0.688306: 50%|##### | 5/10 [00:05<00:05, 1.04s/it][I 2020-09-27 04:53:23,533] Trial 31 finished with value: 0.6888194872886991 and parameters: {'bagging_fraction': 0.994076584593322, 'bagging_freq': 7}. Best is trial 31 with value: 0.6888194872886991.
bagging, val_score: 0.688306: 50%|##### | 5/10 [00:05<00:05, 1.04s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000969 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665112 valid's binary_logloss: 0.689998
Early stopping, best iteration is:
[64] train's binary_logloss: 0.673047 valid's binary_logloss: 0.689362
bagging, val_score: 0.688306: 60%|###### | 6/10 [00:06<00:03, 1.00it/s][I 2020-09-27 04:53:24,435] Trial 32 finished with value: 0.6893623976325989 and parameters: {'bagging_fraction': 0.6317840838824331, 'bagging_freq': 4}. Best is trial 31 with value: 0.6888194872886991.
bagging, val_score: 0.688306: 60%|###### | 6/10 [00:06<00:03, 1.00it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000971 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665587 valid's binary_logloss: 0.689559
[200] train's binary_logloss: 0.646055 valid's binary_logloss: 0.689772
Early stopping, best iteration is:
[147] train's binary_logloss: 0.656183 valid's binary_logloss: 0.689318
bagging, val_score: 0.688306: 70%|####### | 7/10 [00:07<00:03, 1.09s/it][I 2020-09-27 04:53:25,735] Trial 33 finished with value: 0.6893178393179011 and parameters: {'bagging_fraction': 0.6247158545663389, 'bagging_freq': 6}. Best is trial 31 with value: 0.6888194872886991.
bagging, val_score: 0.688306: 70%|####### | 7/10 [00:07<00:03, 1.09s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000971 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665022 valid's binary_logloss: 0.68972
Early stopping, best iteration is:
[60] train's binary_logloss: 0.673867 valid's binary_logloss: 0.689193
bagging, val_score: 0.688306: 80%|######## | 8/10 [00:08<00:02, 1.05s/it][I 2020-09-27 04:53:26,715] Trial 34 finished with value: 0.6891934518367359 and parameters: {'bagging_fraction': 0.8401040019057991, 'bagging_freq': 2}. Best is trial 31 with value: 0.6888194872886991.
bagging, val_score: 0.688306: 80%|######## | 8/10 [00:08<00:02, 1.05s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000949 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665836 valid's binary_logloss: 0.690057
Early stopping, best iteration is:
[26] train's binary_logloss: 0.683137 valid's binary_logloss: 0.69002
bagging, val_score: 0.688306: 90%|######### | 9/10 [00:09<00:00, 1.05it/s][I 2020-09-27 04:53:27,429] Trial 35 finished with value: 0.690019877609649 and parameters: {'bagging_fraction': 0.5293875757229788, 'bagging_freq': 3}. Best is trial 31 with value: 0.6888194872886991.
bagging, val_score: 0.688306: 90%|######### | 9/10 [00:09<00:00, 1.05it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000988 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664989 valid's binary_logloss: 0.690033
Early stopping, best iteration is:
[62] train's binary_logloss: 0.673472 valid's binary_logloss: 0.689508
bagging, val_score: 0.688306: 100%|##########| 10/10 [00:09<00:00, 1.06it/s][I 2020-09-27 04:53:28,340] Trial 36 finished with value: 0.6895082528824183 and parameters: {'bagging_fraction': 0.7844185449233542, 'bagging_freq': 5}. Best is trial 31 with value: 0.6888194872886991.
bagging, val_score: 0.688306: 100%|##########| 10/10 [00:09<00:00, 1.00it/s]
feature_fraction_stage2, val_score: 0.688306: 0%| | 0/6 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000900 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811
[200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036
Early stopping, best iteration is:
[128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306
feature_fraction_stage2, val_score: 0.688306: 17%|#6 | 1/6 [00:01<00:05, 1.08s/it][I 2020-09-27 04:53:29,437] Trial 37 finished with value: 0.6883061543757897 and parameters: {'feature_fraction': 0.516}. Best is trial 37 with value: 0.6883061543757897.
feature_fraction_stage2, val_score: 0.688306: 17%|#6 | 1/6 [00:01<00:05, 1.08s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000788 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664973 valid's binary_logloss: 0.689707
[200] train's binary_logloss: 0.646027 valid's binary_logloss: 0.690321
Early stopping, best iteration is:
[117] train's binary_logloss: 0.661743 valid's binary_logloss: 0.689468
feature_fraction_stage2, val_score: 0.688306: 33%|###3 | 2/6 [00:02<00:04, 1.07s/it][I 2020-09-27 04:53:30,468] Trial 38 finished with value: 0.689467520428726 and parameters: {'feature_fraction': 0.45199999999999996}. Best is trial 37 with value: 0.6883061543757897.
feature_fraction_stage2, val_score: 0.688306: 33%|###3 | 2/6 [00:02<00:04, 1.07s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001140 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.66468 valid's binary_logloss: 0.68972
Early stopping, best iteration is:
[64] train's binary_logloss: 0.672487 valid's binary_logloss: 0.689477
feature_fraction_stage2, val_score: 0.688306: 50%|##### | 3/6 [00:02<00:03, 1.00s/it][I 2020-09-27 04:53:31,330] Trial 39 finished with value: 0.6894774349755185 and parameters: {'feature_fraction': 0.58}. Best is trial 37 with value: 0.6883061543757897.
feature_fraction_stage2, val_score: 0.688306: 50%|##### | 3/6 [00:02<00:03, 1.00s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000843 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.66569 valid's binary_logloss: 0.688902
Early stopping, best iteration is:
[65] train's binary_logloss: 0.673083 valid's binary_logloss: 0.688613
feature_fraction_stage2, val_score: 0.688306: 67%|######6 | 4/6 [00:03<00:01, 1.06it/s][I 2020-09-27 04:53:32,142] Trial 40 finished with value: 0.6886133673965851 and parameters: {'feature_fraction': 0.42}. Best is trial 37 with value: 0.6883061543757897.
feature_fraction_stage2, val_score: 0.688306: 67%|######6 | 4/6 [00:03<00:01, 1.06it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000999 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664812 valid's binary_logloss: 0.688514
Early stopping, best iteration is:
[92] train's binary_logloss: 0.66641 valid's binary_logloss: 0.688368
feature_fraction_stage2, val_score: 0.688306: 83%|########3 | 5/6 [00:04<00:00, 1.05it/s][I 2020-09-27 04:53:33,108] Trial 41 finished with value: 0.6883683592481646 and parameters: {'feature_fraction': 0.5479999999999999}. Best is trial 37 with value: 0.6883061543757897.
feature_fraction_stage2, val_score: 0.688306: 83%|########3 | 5/6 [00:04<00:00, 1.05it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000975 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811
[200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036
Early stopping, best iteration is:
[128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306
feature_fraction_stage2, val_score: 0.688306: 100%|##########| 6/6 [00:05<00:00, 1.01it/s][I 2020-09-27 04:53:34,199] Trial 42 finished with value: 0.6883061543757897 and parameters: {'feature_fraction': 0.484}. Best is trial 37 with value: 0.6883061543757897.
feature_fraction_stage2, val_score: 0.688306: 100%|##########| 6/6 [00:05<00:00, 1.03it/s]
regularization_factors, val_score: 0.688306: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000960 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.66526 valid's binary_logloss: 0.689325
Early stopping, best iteration is:
[74] train's binary_logloss: 0.670714 valid's binary_logloss: 0.688927
regularization_factors, val_score: 0.688306: 5%|5 | 1/20 [00:00<00:16, 1.13it/s][I 2020-09-27 04:53:35,096] Trial 43 finished with value: 0.6889268680203512 and parameters: {'lambda_l1': 3.489683036490915e-05, 'lambda_l2': 0.03596499685458891}. Best is trial 43 with value: 0.6889268680203512.
regularization_factors, val_score: 0.688306: 5%|5 | 1/20 [00:00<00:16, 1.13it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000943 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.668189 valid's binary_logloss: 0.688592
Early stopping, best iteration is:
[76] train's binary_logloss: 0.672479 valid's binary_logloss: 0.688382
regularization_factors, val_score: 0.688306: 10%|# | 2/20 [00:01<00:16, 1.11it/s][I 2020-09-27 04:53:36,045] Trial 44 finished with value: 0.6883820148603388 and parameters: {'lambda_l1': 4.2910298459529574, 'lambda_l2': 4.4841181697512024e-08}. Best is trial 44 with value: 0.6883820148603388.
regularization_factors, val_score: 0.688306: 10%|# | 2/20 [00:01<00:16, 1.11it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001239 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811
[200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036
Early stopping, best iteration is:
[128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306
regularization_factors, val_score: 0.688306: 15%|#5 | 3/20 [00:02<00:16, 1.04it/s][I 2020-09-27 04:53:37,146] Trial 45 finished with value: 0.6883061543637848 and parameters: {'lambda_l1': 1.6123080908302226e-08, 'lambda_l2': 1.1923474224643993e-07}. Best is trial 45 with value: 0.6883061543637848.
regularization_factors, val_score: 0.688306: 15%|#5 | 3/20 [00:02<00:16, 1.04it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001032 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811
[200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036
Early stopping, best iteration is:
[128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306
regularization_factors, val_score: 0.688306: 20%|## | 4/20 [00:04<00:16, 1.01s/it][I 2020-09-27 04:53:38,271] Trial 46 finished with value: 0.6883061543726559 and parameters: {'lambda_l1': 1.4997584226077242e-08, 'lambda_l2': 1.9309033302845346e-08}. Best is trial 45 with value: 0.6883061543637848.
regularization_factors, val_score: 0.688306: 20%|## | 4/20 [00:04<00:16, 1.01s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000984 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811
[200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036
Early stopping, best iteration is:
[128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306
regularization_factors, val_score: 0.688306: 25%|##5 | 5/20 [00:05<00:15, 1.05s/it][I 2020-09-27 04:53:39,412] Trial 47 finished with value: 0.6883061543725224 and parameters: {'lambda_l1': 2.1537994864162705e-08, 'lambda_l2': 1.7013032790342447e-08}. Best is trial 45 with value: 0.6883061543637848.
regularization_factors, val_score: 0.688306: 25%|##5 | 5/20 [00:05<00:15, 1.05s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000950 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811
[200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036
Early stopping, best iteration is:
[128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306
regularization_factors, val_score: 0.688306: 30%|### | 6/20 [00:06<00:14, 1.06s/it][I 2020-09-27 04:53:40,511] Trial 48 finished with value: 0.6883061543697307 and parameters: {'lambda_l1': 4.5182334372625207e-08, 'lambda_l2': 1.1900726497967697e-08}. Best is trial 45 with value: 0.6883061543637848.
regularization_factors, val_score: 0.688306: 30%|### | 6/20 [00:06<00:14, 1.06s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000959 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811
[200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036
Early stopping, best iteration is:
[128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306
regularization_factors, val_score: 0.688306: 35%|###5 | 7/20 [00:07<00:13, 1.07s/it][I 2020-09-27 04:53:41,604] Trial 49 finished with value: 0.6883061543737345 and parameters: {'lambda_l1': 1.214350265927941e-08, 'lambda_l2': 1.243915863528419e-08}. Best is trial 45 with value: 0.6883061543637848.
regularization_factors, val_score: 0.688306: 35%|###5 | 7/20 [00:07<00:13, 1.07s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000942 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811
[200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036
Early stopping, best iteration is:
[128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306
regularization_factors, val_score: 0.688306: 40%|#### | 8/20 [00:08<00:12, 1.08s/it][I 2020-09-27 04:53:42,706] Trial 50 finished with value: 0.6883061543740199 and parameters: {'lambda_l1': 1.0018700830284507e-08, 'lambda_l2': 1.1810016327451072e-08}. Best is trial 45 with value: 0.6883061543637848.
regularization_factors, val_score: 0.688306: 40%|#### | 8/20 [00:08<00:12, 1.08s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000961 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811
[200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036
Early stopping, best iteration is:
[128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306
regularization_factors, val_score: 0.688306: 45%|####5 | 9/20 [00:09<00:12, 1.10s/it][I 2020-09-27 04:53:43,836] Trial 51 finished with value: 0.6883061543739415 and parameters: {'lambda_l1': 1.034818635491755e-08, 'lambda_l2': 1.1390198269989985e-08}. Best is trial 45 with value: 0.6883061543637848.
regularization_factors, val_score: 0.688306: 45%|####5 | 9/20 [00:09<00:12, 1.10s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000968 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811
[200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036
Early stopping, best iteration is:
[128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306
regularization_factors, val_score: 0.688306: 50%|##### | 10/20 [00:10<00:11, 1.12s/it][I 2020-09-27 04:53:45,004] Trial 52 finished with value: 0.6883061543730142 and parameters: {'lambda_l1': 1.67980865791993e-08, 'lambda_l2': 1.4273245882473324e-08}. Best is trial 45 with value: 0.6883061543637848.
regularization_factors, val_score: 0.688306: 50%|##### | 10/20 [00:10<00:11, 1.12s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000947 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811
[200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036
Early stopping, best iteration is:
[128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306
regularization_factors, val_score: 0.688306: 55%|#####5 | 11/20 [00:11<00:10, 1.11s/it][I 2020-09-27 04:53:46,102] Trial 53 finished with value: 0.6883061543727123 and parameters: {'lambda_l1': 1.54156161065481e-08, 'lambda_l2': 1.961580255914531e-08}. Best is trial 45 with value: 0.6883061543637848.
regularization_factors, val_score: 0.688306: 55%|#####5 | 11/20 [00:11<00:10, 1.11s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001597 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811
[200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036
Early stopping, best iteration is:
[128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306
regularization_factors, val_score: 0.688306: 60%|###### | 12/20 [00:12<00:08, 1.10s/it][I 2020-09-27 04:53:47,171] Trial 54 finished with value: 0.688306154369646 and parameters: {'lambda_l1': 2.6426720558168387e-08, 'lambda_l2': 3.5118264244271155e-08}. Best is trial 45 with value: 0.6883061543637848.
regularization_factors, val_score: 0.688306: 60%|###### | 12/20 [00:12<00:08, 1.10s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000949 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811
[200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036
Early stopping, best iteration is:
[128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306
regularization_factors, val_score: 0.688306: 65%|######5 | 13/20 [00:14<00:07, 1.10s/it][I 2020-09-27 04:53:48,280] Trial 55 finished with value: 0.6883061543276774 and parameters: {'lambda_l1': 6.8998257219367e-08, 'lambda_l2': 5.65558107194084e-07}. Best is trial 55 with value: 0.6883061543276774.
regularization_factors, val_score: 0.688306: 65%|######5 | 13/20 [00:14<00:07, 1.10s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001012 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811
[200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036
Early stopping, best iteration is:
[128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306
regularization_factors, val_score: 0.688306: 70%|####### | 14/20 [00:15<00:06, 1.13s/it][I 2020-09-27 04:53:49,474] Trial 56 finished with value: 0.6883061540275088 and parameters: {'lambda_l1': 7.521323057192409e-07, 'lambda_l2': 3.7320052285004596e-06}. Best is trial 56 with value: 0.6883061540275088.
regularization_factors, val_score: 0.688306: 70%|####### | 14/20 [00:15<00:06, 1.13s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000962 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811
[200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036
Early stopping, best iteration is:
[128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306
regularization_factors, val_score: 0.688306: 75%|#######5 | 15/20 [00:16<00:05, 1.13s/it][I 2020-09-27 04:53:50,589] Trial 57 finished with value: 0.6883061539600913 and parameters: {'lambda_l1': 1.1451365202695885e-06, 'lambda_l2': 4.1525999441678045e-06}. Best is trial 57 with value: 0.6883061539600913.
regularization_factors, val_score: 0.688306: 75%|#######5 | 15/20 [00:16<00:05, 1.13s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000947 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811
[200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036
Early stopping, best iteration is:
[128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306
regularization_factors, val_score: 0.688306: 80%|######## | 16/20 [00:17<00:04, 1.12s/it][I 2020-09-27 04:53:51,694] Trial 58 finished with value: 0.6883061537046381 and parameters: {'lambda_l1': 1.87654882489576e-06, 'lambda_l2': 6.539895147651026e-06}. Best is trial 58 with value: 0.6883061537046381.
regularization_factors, val_score: 0.688306: 80%|######## | 16/20 [00:17<00:04, 1.12s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000888 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811
[200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036
Early stopping, best iteration is:
[128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306
regularization_factors, val_score: 0.688306: 85%|########5 | 17/20 [00:18<00:03, 1.11s/it][I 2020-09-27 04:53:52,787] Trial 59 finished with value: 0.6883061539934588 and parameters: {'lambda_l1': 1.7458283390546473e-06, 'lambda_l2': 2.9537260784896182e-06}. Best is trial 58 with value: 0.6883061537046381.
regularization_factors, val_score: 0.688306: 85%|########5 | 17/20 [00:18<00:03, 1.11s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000908 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811
[200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036
Early stopping, best iteration is:
[128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306
regularization_factors, val_score: 0.688306: 90%|######### | 18/20 [00:19<00:02, 1.11s/it][I 2020-09-27 04:53:53,910] Trial 60 finished with value: 0.6883061538065175 and parameters: {'lambda_l1': 2.402166405130976e-06, 'lambda_l2': 4.623828082213961e-06}. Best is trial 58 with value: 0.6883061537046381.
regularization_factors, val_score: 0.688306: 90%|######### | 18/20 [00:19<00:02, 1.11s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010324 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811
[200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036
Early stopping, best iteration is:
[128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306
regularization_factors, val_score: 0.688306: 95%|#########5| 19/20 [00:20<00:01, 1.12s/it][I 2020-09-27 04:53:55,054] Trial 61 finished with value: 0.6883061537873091 and parameters: {'lambda_l1': 2.6852977733488266e-06, 'lambda_l2': 4.514170857027213e-06}. Best is trial 58 with value: 0.6883061537046381.
regularization_factors, val_score: 0.688306: 95%|#########5| 19/20 [00:20<00:01, 1.12s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012564 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811
[200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036
Early stopping, best iteration is:
[128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306
regularization_factors, val_score: 0.688306: 100%|##########| 20/20 [00:21<00:00, 1.10s/it][I 2020-09-27 04:53:56,095] Trial 62 finished with value: 0.6883061538017431 and parameters: {'lambda_l1': 2.099566024643849e-06, 'lambda_l2': 5.0450298575946105e-06}. Best is trial 58 with value: 0.6883061537046381.
regularization_factors, val_score: 0.688306: 100%|##########| 20/20 [00:21<00:00, 1.09s/it]
min_data_in_leaf, val_score: 0.688306: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000964 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665071 valid's binary_logloss: 0.68918
Early stopping, best iteration is:
[66] train's binary_logloss: 0.672155 valid's binary_logloss: 0.688846
min_data_in_leaf, val_score: 0.688306: 20%|## | 1/5 [00:00<00:03, 1.20it/s][I 2020-09-27 04:53:56,941] Trial 63 finished with value: 0.6888461130563782 and parameters: {'min_child_samples': 10}. Best is trial 63 with value: 0.6888461130563782.
min_data_in_leaf, val_score: 0.688306: 20%|## | 1/5 [00:00<00:03, 1.20it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001044 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.666223 valid's binary_logloss: 0.689352
Early stopping, best iteration is:
[63] train's binary_logloss: 0.673794 valid's binary_logloss: 0.688948
min_data_in_leaf, val_score: 0.688306: 40%|#### | 2/5 [00:01<00:02, 1.20it/s][I 2020-09-27 04:53:57,783] Trial 64 finished with value: 0.6889480364726907 and parameters: {'min_child_samples': 100}. Best is trial 63 with value: 0.6888461130563782.
min_data_in_leaf, val_score: 0.688306: 40%|#### | 2/5 [00:01<00:02, 1.20it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000895 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.6653 valid's binary_logloss: 0.688213
Early stopping, best iteration is:
[98] train's binary_logloss: 0.665652 valid's binary_logloss: 0.688167
min_data_in_leaf, val_score: 0.688167: 60%|###### | 3/5 [00:02<00:01, 1.14it/s][I 2020-09-27 04:53:58,761] Trial 65 finished with value: 0.6881665249243496 and parameters: {'min_child_samples': 25}. Best is trial 65 with value: 0.6881665249243496.
min_data_in_leaf, val_score: 0.688167: 60%|###### | 3/5 [00:02<00:01, 1.14it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000967 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664957 valid's binary_logloss: 0.688576
Early stopping, best iteration is:
[99] train's binary_logloss: 0.665189 valid's binary_logloss: 0.688543
min_data_in_leaf, val_score: 0.688167: 80%|######## | 4/5 [00:03<00:00, 1.09it/s][I 2020-09-27 04:53:59,775] Trial 66 finished with value: 0.6885432542311284 and parameters: {'min_child_samples': 5}. Best is trial 65 with value: 0.6881665249243496.
min_data_in_leaf, val_score: 0.688167: 80%|######## | 4/5 [00:03<00:00, 1.09it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001196 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288
[LightGBM] [Info] Start training from score 0.009288
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665534 valid's binary_logloss: 0.689168
Early stopping, best iteration is:
[86] train's binary_logloss: 0.668451 valid's binary_logloss: 0.688864
min_data_in_leaf, val_score: 0.688167: 100%|##########| 5/5 [00:04<00:00, 1.07it/s][I 2020-09-27 04:54:00,741] Trial 67 finished with value: 0.6888639866159642 and parameters: {'min_child_samples': 50}. Best is trial 65 with value: 0.6881665249243496.
min_data_in_leaf, val_score: 0.688167: 100%|##########| 5/5 [00:04<00:00, 1.08it/s]
Fold : 3
[I 2020-09-27 04:54:00,831] A new study created in memory with name: no-name-2a0ad92a-db0c-44a0-a535-c4566ac6f3da
feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010784 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.66381 valid's binary_logloss: 0.689338
Early stopping, best iteration is:
[57] train's binary_logloss: 0.673532 valid's binary_logloss: 0.68897
feature_fraction, val_score: 0.688970: 14%|#4 | 1/7 [00:00<00:05, 1.01it/s][I 2020-09-27 04:54:01,833] Trial 0 finished with value: 0.6889700252608528 and parameters: {'feature_fraction': 0.7}. Best is trial 0 with value: 0.6889700252608528.
feature_fraction, val_score: 0.688970: 14%|#4 | 1/7 [00:00<00:05, 1.01it/s][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001634 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.66267 valid's binary_logloss: 0.689828
Early stopping, best iteration is:
[65] train's binary_logloss: 0.670714 valid's binary_logloss: 0.689441
feature_fraction, val_score: 0.688970: 29%|##8 | 2/7 [00:01<00:04, 1.04it/s][I 2020-09-27 04:54:02,732] Trial 1 finished with value: 0.6894408688583004 and parameters: {'feature_fraction': 1.0}. Best is trial 0 with value: 0.6889700252608528.
feature_fraction, val_score: 0.688970: 29%|##8 | 2/7 [00:01<00:04, 1.04it/s][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000679 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.666188 valid's binary_logloss: 0.689964
[200] train's binary_logloss: 0.648164 valid's binary_logloss: 0.690421
Early stopping, best iteration is:
[146] train's binary_logloss: 0.657493 valid's binary_logloss: 0.689838
feature_fraction, val_score: 0.688970: 43%|####2 | 3/7 [00:02<00:04, 1.00s/it][I 2020-09-27 04:54:03,817] Trial 2 finished with value: 0.6898376157358569 and parameters: {'feature_fraction': 0.4}. Best is trial 0 with value: 0.6889700252608528.
feature_fraction, val_score: 0.688970: 43%|####2 | 3/7 [00:02<00:04, 1.00s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008176 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664448 valid's binary_logloss: 0.690083
Early stopping, best iteration is:
[61] train's binary_logloss: 0.673008 valid's binary_logloss: 0.68947
feature_fraction, val_score: 0.688970: 57%|#####7 | 4/7 [00:03<00:02, 1.08it/s][I 2020-09-27 04:54:04,577] Trial 3 finished with value: 0.6894700918769894 and parameters: {'feature_fraction': 0.6}. Best is trial 0 with value: 0.6889700252608528.
feature_fraction, val_score: 0.688970: 57%|#####7 | 4/7 [00:03<00:02, 1.08it/s][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001461 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.6633 valid's binary_logloss: 0.68927
[200] train's binary_logloss: 0.642614 valid's binary_logloss: 0.689045
Early stopping, best iteration is:
[194] train's binary_logloss: 0.643867 valid's binary_logloss: 0.688894
feature_fraction, val_score: 0.688894: 71%|#######1 | 5/7 [00:05<00:02, 1.10s/it][I 2020-09-27 04:54:06,090] Trial 4 finished with value: 0.6888944409719155 and parameters: {'feature_fraction': 0.8}. Best is trial 4 with value: 0.6888944409719155.
feature_fraction, val_score: 0.688894: 71%|#######1 | 5/7 [00:05<00:02, 1.10s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001711 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.662969 valid's binary_logloss: 0.689906
Early stopping, best iteration is:
[58] train's binary_logloss: 0.672958 valid's binary_logloss: 0.689479
feature_fraction, val_score: 0.688894: 86%|########5 | 6/7 [00:06<00:01, 1.04s/it][I 2020-09-27 04:54:06,969] Trial 5 finished with value: 0.6894794839977714 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 4 with value: 0.6888944409719155.
feature_fraction, val_score: 0.688894: 86%|########5 | 6/7 [00:06<00:01, 1.04s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001143 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664948 valid's binary_logloss: 0.689939
Early stopping, best iteration is:
[71] train's binary_logloss: 0.671105 valid's binary_logloss: 0.689412
feature_fraction, val_score: 0.688894: 100%|##########| 7/7 [00:06<00:00, 1.02it/s][I 2020-09-27 04:54:07,829] Trial 6 finished with value: 0.689411673782528 and parameters: {'feature_fraction': 0.5}. Best is trial 4 with value: 0.6888944409719155.
feature_fraction, val_score: 0.688894: 100%|##########| 7/7 [00:06<00:00, 1.00it/s]
num_leaves, val_score: 0.688894: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016008 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.545996 valid's binary_logloss: 0.696346
Early stopping, best iteration is:
[20] train's binary_logloss: 0.650465 valid's binary_logloss: 0.691113
num_leaves, val_score: 0.688894: 5%|5 | 1/20 [00:01<00:25, 1.32s/it][I 2020-09-27 04:54:09,164] Trial 7 finished with value: 0.6911130787415561 and parameters: {'num_leaves': 219}. Best is trial 7 with value: 0.6911130787415561.
num_leaves, val_score: 0.688894: 5%|5 | 1/20 [00:01<00:25, 1.32s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001458 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.583507 valid's binary_logloss: 0.694216
Early stopping, best iteration is:
[18] train's binary_logloss: 0.66415 valid's binary_logloss: 0.691315
num_leaves, val_score: 0.688894: 10%|# | 2/20 [00:02<00:23, 1.28s/it][I 2020-09-27 04:54:10,338] Trial 8 finished with value: 0.6913152010105598 and parameters: {'num_leaves': 152}. Best is trial 7 with value: 0.6911130787415561.
num_leaves, val_score: 0.688894: 10%|# | 2/20 [00:02<00:23, 1.28s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001515 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.600147 valid's binary_logloss: 0.692792
Early stopping, best iteration is:
[45] train's binary_logloss: 0.642663 valid's binary_logloss: 0.690464
num_leaves, val_score: 0.688894: 15%|#5 | 3/20 [00:04<00:23, 1.39s/it][I 2020-09-27 04:54:11,983] Trial 9 finished with value: 0.690463587315969 and parameters: {'num_leaves': 123}. Best is trial 9 with value: 0.690463587315969.
num_leaves, val_score: 0.688894: 15%|#5 | 3/20 [00:04<00:23, 1.39s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001721 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.676573 valid's binary_logloss: 0.689403
Early stopping, best iteration is:
[82] train's binary_logloss: 0.678598 valid's binary_logloss: 0.689178
num_leaves, val_score: 0.688894: 20%|## | 4/20 [00:06<00:25, 1.61s/it][I 2020-09-27 04:54:14,106] Trial 10 finished with value: 0.6891784496510537 and parameters: {'num_leaves': 15}. Best is trial 10 with value: 0.6891784496510537.
num_leaves, val_score: 0.688894: 20%|## | 4/20 [00:06<00:25, 1.61s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012855 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689973 valid's binary_logloss: 0.691198
[200] train's binary_logloss: 0.688794 valid's binary_logloss: 0.690538
[300] train's binary_logloss: 0.68806 valid's binary_logloss: 0.690049
[400] train's binary_logloss: 0.687537 valid's binary_logloss: 0.689777
[500] train's binary_logloss: 0.687144 valid's binary_logloss: 0.689665
[600] train's binary_logloss: 0.686839 valid's binary_logloss: 0.689559
[700] train's binary_logloss: 0.686592 valid's binary_logloss: 0.689472
[800] train's binary_logloss: 0.686384 valid's binary_logloss: 0.689469
[900] train's binary_logloss: 0.686205 valid's binary_logloss: 0.689458
Early stopping, best iteration is:
[877] train's binary_logloss: 0.686244 valid's binary_logloss: 0.68944
num_leaves, val_score: 0.688894: 25%|##5 | 5/20 [00:09<00:31, 2.13s/it][I 2020-09-27 04:54:17,448] Trial 11 finished with value: 0.689440488741354 and parameters: {'num_leaves': 2}. Best is trial 10 with value: 0.6891784496510537.
num_leaves, val_score: 0.688894: 25%|##5 | 5/20 [00:09<00:31, 2.13s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001640 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.531631 valid's binary_logloss: 0.696171
Early stopping, best iteration is:
[9] train's binary_logloss: 0.669502 valid's binary_logloss: 0.691204
num_leaves, val_score: 0.688894: 30%|### | 6/20 [00:11<00:27, 1.95s/it][I 2020-09-27 04:54:18,964] Trial 12 finished with value: 0.6912044508576668 and parameters: {'num_leaves': 252}. Best is trial 10 with value: 0.6891784496510537.
num_leaves, val_score: 0.688894: 30%|### | 6/20 [00:11<00:27, 1.95s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001636 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.626032 valid's binary_logloss: 0.690485
Early stopping, best iteration is:
[80] train's binary_logloss: 0.636438 valid's binary_logloss: 0.689817
num_leaves, val_score: 0.688894: 35%|###5 | 7/20 [00:12<00:22, 1.75s/it][I 2020-09-27 04:54:20,263] Trial 13 finished with value: 0.6898168565458592 and parameters: {'num_leaves': 83}. Best is trial 10 with value: 0.6891784496510537.
num_leaves, val_score: 0.688894: 35%|###5 | 7/20 [00:12<00:22, 1.75s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010079 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.559692 valid's binary_logloss: 0.696329
Early stopping, best iteration is:
[18] train's binary_logloss: 0.657719 valid's binary_logloss: 0.691223
num_leaves, val_score: 0.688894: 40%|#### | 8/20 [00:13<00:18, 1.58s/it][I 2020-09-27 04:54:21,443] Trial 14 finished with value: 0.6912227645026857 and parameters: {'num_leaves': 197}. Best is trial 10 with value: 0.6891784496510537.
num_leaves, val_score: 0.688894: 40%|#### | 8/20 [00:13<00:18, 1.58s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001528 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.643665 valid's binary_logloss: 0.691255
Early stopping, best iteration is:
[27] train's binary_logloss: 0.67467 valid's binary_logloss: 0.689948
num_leaves, val_score: 0.688894: 45%|####5 | 9/20 [00:14<00:14, 1.36s/it][I 2020-09-27 04:54:22,296] Trial 15 finished with value: 0.6899483024551792 and parameters: {'num_leaves': 57}. Best is trial 10 with value: 0.6891784496510537.
num_leaves, val_score: 0.688894: 45%|####5 | 9/20 [00:14<00:14, 1.36s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001941 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.5757 valid's binary_logloss: 0.695653
Early stopping, best iteration is:
[17] train's binary_logloss: 0.663745 valid's binary_logloss: 0.691071
num_leaves, val_score: 0.688894: 50%|##### | 10/20 [00:15<00:13, 1.32s/it][I 2020-09-27 04:54:23,534] Trial 16 finished with value: 0.6910709930215578 and parameters: {'num_leaves': 163}. Best is trial 10 with value: 0.6891784496510537.
num_leaves, val_score: 0.688894: 50%|##### | 10/20 [00:15<00:13, 1.32s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008405 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.642489 valid's binary_logloss: 0.690471
Early stopping, best iteration is:
[45] train's binary_logloss: 0.665549 valid's binary_logloss: 0.689944
num_leaves, val_score: 0.688894: 55%|#####5 | 11/20 [00:16<00:10, 1.20s/it][I 2020-09-27 04:54:24,436] Trial 17 finished with value: 0.6899435817261527 and parameters: {'num_leaves': 58}. Best is trial 10 with value: 0.6891784496510537.
num_leaves, val_score: 0.688894: 55%|#####5 | 11/20 [00:16<00:10, 1.20s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002315 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.536859 valid's binary_logloss: 0.696258
Early stopping, best iteration is:
[24] train's binary_logloss: 0.640234 valid's binary_logloss: 0.690993
num_leaves, val_score: 0.688894: 60%|###### | 12/20 [00:18<00:10, 1.32s/it][I 2020-09-27 04:54:26,057] Trial 18 finished with value: 0.6909927461103317 and parameters: {'num_leaves': 239}. Best is trial 10 with value: 0.6891784496510537.
num_leaves, val_score: 0.688894: 60%|###### | 12/20 [00:18<00:10, 1.32s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001510 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.618426 valid's binary_logloss: 0.69242
Early stopping, best iteration is:
[22] train's binary_logloss: 0.669782 valid's binary_logloss: 0.69036
num_leaves, val_score: 0.688894: 65%|######5 | 13/20 [00:19<00:08, 1.22s/it][I 2020-09-27 04:54:27,037] Trial 19 finished with value: 0.6903597022330396 and parameters: {'num_leaves': 95}. Best is trial 10 with value: 0.6891784496510537.
num_leaves, val_score: 0.688894: 65%|######5 | 13/20 [00:19<00:08, 1.22s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.014205 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.568032 valid's binary_logloss: 0.694802
Early stopping, best iteration is:
[20] train's binary_logloss: 0.656477 valid's binary_logloss: 0.690174
num_leaves, val_score: 0.688894: 70%|####### | 14/20 [00:20<00:07, 1.20s/it][I 2020-09-27 04:54:28,190] Trial 20 finished with value: 0.690174257788127 and parameters: {'num_leaves': 182}. Best is trial 10 with value: 0.6891784496510537.
num_leaves, val_score: 0.688894: 70%|####### | 14/20 [00:20<00:07, 1.20s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.006271 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681177 valid's binary_logloss: 0.689414
[200] train's binary_logloss: 0.674276 valid's binary_logloss: 0.689006
Early stopping, best iteration is:
[194] train's binary_logloss: 0.674612 valid's binary_logloss: 0.688884
num_leaves, val_score: 0.688884: 75%|#######5 | 15/20 [00:21<00:06, 1.24s/it][I 2020-09-27 04:54:29,535] Trial 21 finished with value: 0.6888838019748574 and parameters: {'num_leaves': 10}. Best is trial 21 with value: 0.6888838019748574.
num_leaves, val_score: 0.688884: 75%|#######5 | 15/20 [00:21<00:06, 1.24s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001987 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663848 valid's binary_logloss: 0.689368
Early stopping, best iteration is:
[90] train's binary_logloss: 0.666139 valid's binary_logloss: 0.689192
num_leaves, val_score: 0.688884: 80%|######## | 16/20 [00:22<00:04, 1.19s/it][I 2020-09-27 04:54:30,589] Trial 22 finished with value: 0.6891919324492329 and parameters: {'num_leaves': 30}. Best is trial 21 with value: 0.6888838019748574.
num_leaves, val_score: 0.688884: 80%|######## | 16/20 [00:22<00:04, 1.19s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009391 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.605617 valid's binary_logloss: 0.693796
Early stopping, best iteration is:
[22] train's binary_logloss: 0.666098 valid's binary_logloss: 0.690873
num_leaves, val_score: 0.688884: 85%|########5 | 17/20 [00:23<00:03, 1.13s/it][I 2020-09-27 04:54:31,594] Trial 23 finished with value: 0.6908732323148996 and parameters: {'num_leaves': 114}. Best is trial 21 with value: 0.6888838019748574.
num_leaves, val_score: 0.688884: 85%|########5 | 17/20 [00:23<00:03, 1.13s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001645 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.655426 valid's binary_logloss: 0.689235
[200] train's binary_logloss: 0.628962 valid's binary_logloss: 0.691293
Early stopping, best iteration is:
[106] train's binary_logloss: 0.653726 valid's binary_logloss: 0.689075
num_leaves, val_score: 0.688884: 90%|######### | 18/20 [00:24<00:02, 1.14s/it][I 2020-09-27 04:54:32,740] Trial 24 finished with value: 0.6890748065383784 and parameters: {'num_leaves': 41}. Best is trial 21 with value: 0.6888838019748574.
num_leaves, val_score: 0.688884: 90%|######### | 18/20 [00:24<00:02, 1.14s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011385 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687114 valid's binary_logloss: 0.690103
[200] train's binary_logloss: 0.684519 valid's binary_logloss: 0.689685
[300] train's binary_logloss: 0.682489 valid's binary_logloss: 0.68949
[400] train's binary_logloss: 0.680652 valid's binary_logloss: 0.689206
[500] train's binary_logloss: 0.678959 valid's binary_logloss: 0.689203
Early stopping, best iteration is:
[432] train's binary_logloss: 0.680085 valid's binary_logloss: 0.689077
num_leaves, val_score: 0.688884: 95%|#########5| 19/20 [00:26<00:01, 1.37s/it][I 2020-09-27 04:54:34,648] Trial 25 finished with value: 0.6890769719746892 and parameters: {'num_leaves': 4}. Best is trial 21 with value: 0.6888838019748574.
num_leaves, val_score: 0.688884: 95%|#########5| 19/20 [00:26<00:01, 1.37s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002085 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.624994 valid's binary_logloss: 0.690549
Early stopping, best iteration is:
[49] train's binary_logloss: 0.653531 valid's binary_logloss: 0.689478
num_leaves, val_score: 0.688884: 100%|##########| 20/20 [00:27<00:00, 1.30s/it][I 2020-09-27 04:54:35,793] Trial 26 finished with value: 0.6894778249561458 and parameters: {'num_leaves': 84}. Best is trial 21 with value: 0.6888838019748574.
num_leaves, val_score: 0.688884: 100%|##########| 20/20 [00:27<00:00, 1.40s/it]
bagging, val_score: 0.688884: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013914 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681106 valid's binary_logloss: 0.689431
[200] train's binary_logloss: 0.674446 valid's binary_logloss: 0.689212
[300] train's binary_logloss: 0.668203 valid's binary_logloss: 0.689716
Early stopping, best iteration is:
[201] train's binary_logloss: 0.674354 valid's binary_logloss: 0.6892
bagging, val_score: 0.688884: 10%|# | 1/10 [00:01<00:11, 1.28s/it][I 2020-09-27 04:54:37,087] Trial 27 finished with value: 0.6891997209813286 and parameters: {'bagging_fraction': 0.5047216967290817, 'bagging_freq': 2}. Best is trial 27 with value: 0.6891997209813286.
bagging, val_score: 0.688884: 10%|# | 1/10 [00:01<00:11, 1.28s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001692 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681088 valid's binary_logloss: 0.689391
[200] train's binary_logloss: 0.674086 valid's binary_logloss: 0.689322
Early stopping, best iteration is:
[195] train's binary_logloss: 0.67442 valid's binary_logloss: 0.689229
bagging, val_score: 0.688884: 20%|## | 2/10 [00:02<00:10, 1.32s/it][I 2020-09-27 04:54:38,476] Trial 28 finished with value: 0.6892290270817055 and parameters: {'bagging_fraction': 0.9953814685875817, 'bagging_freq': 7}. Best is trial 27 with value: 0.6891997209813286.
bagging, val_score: 0.688884: 20%|## | 2/10 [00:02<00:10, 1.32s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.014930 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681054 valid's binary_logloss: 0.689888
[200] train's binary_logloss: 0.674088 valid's binary_logloss: 0.689474
[300] train's binary_logloss: 0.667589 valid's binary_logloss: 0.689818
Early stopping, best iteration is:
[201] train's binary_logloss: 0.674022 valid's binary_logloss: 0.689449
bagging, val_score: 0.688884: 30%|### | 3/10 [00:04<00:09, 1.34s/it][I 2020-09-27 04:54:39,878] Trial 29 finished with value: 0.6894491770687096 and parameters: {'bagging_fraction': 0.9657183733734178, 'bagging_freq': 7}. Best is trial 27 with value: 0.6891997209813286.
bagging, val_score: 0.688884: 30%|### | 3/10 [00:04<00:09, 1.34s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001797 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681277 valid's binary_logloss: 0.689585
[200] train's binary_logloss: 0.674417 valid's binary_logloss: 0.689825
Early stopping, best iteration is:
[161] train's binary_logloss: 0.676999 valid's binary_logloss: 0.689344
bagging, val_score: 0.688884: 40%|#### | 4/10 [00:05<00:07, 1.32s/it][I 2020-09-27 04:54:41,134] Trial 30 finished with value: 0.6893441175783354 and parameters: {'bagging_fraction': 0.4138320408319066, 'bagging_freq': 1}. Best is trial 27 with value: 0.6891997209813286.
bagging, val_score: 0.688884: 40%|#### | 4/10 [00:05<00:07, 1.32s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008705 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681066 valid's binary_logloss: 0.688712
[200] train's binary_logloss: 0.674172 valid's binary_logloss: 0.688921
Early stopping, best iteration is:
[121] train's binary_logloss: 0.679523 valid's binary_logloss: 0.688425
bagging, val_score: 0.688425: 50%|##### | 5/10 [00:06<00:06, 1.23s/it][I 2020-09-27 04:54:42,157] Trial 31 finished with value: 0.6884249102782825 and parameters: {'bagging_fraction': 0.7249814340678598, 'bagging_freq': 4}. Best is trial 31 with value: 0.6884249102782825.
bagging, val_score: 0.688425: 50%|##### | 5/10 [00:06<00:06, 1.23s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001643 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681083 valid's binary_logloss: 0.689228
[200] train's binary_logloss: 0.674127 valid's binary_logloss: 0.688871
[300] train's binary_logloss: 0.667696 valid's binary_logloss: 0.688939
Early stopping, best iteration is:
[247] train's binary_logloss: 0.671103 valid's binary_logloss: 0.688664
bagging, val_score: 0.688425: 60%|###### | 6/10 [00:07<00:05, 1.31s/it][I 2020-09-27 04:54:43,668] Trial 32 finished with value: 0.6886636187874674 and parameters: {'bagging_fraction': 0.7386084209140283, 'bagging_freq': 4}. Best is trial 31 with value: 0.6884249102782825.
bagging, val_score: 0.688425: 60%|###### | 6/10 [00:07<00:05, 1.31s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001677 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681024 valid's binary_logloss: 0.689172
[200] train's binary_logloss: 0.67391 valid's binary_logloss: 0.689047
Early stopping, best iteration is:
[125] train's binary_logloss: 0.67913 valid's binary_logloss: 0.688666
bagging, val_score: 0.688425: 70%|####### | 7/10 [00:08<00:03, 1.24s/it][I 2020-09-27 04:54:44,744] Trial 33 finished with value: 0.6886664091503993 and parameters: {'bagging_fraction': 0.765888231588719, 'bagging_freq': 4}. Best is trial 31 with value: 0.6884249102782825.
bagging, val_score: 0.688425: 70%|####### | 7/10 [00:08<00:03, 1.24s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009276 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681076 valid's binary_logloss: 0.689153
[200] train's binary_logloss: 0.674066 valid's binary_logloss: 0.688436
[300] train's binary_logloss: 0.667548 valid's binary_logloss: 0.688896
Early stopping, best iteration is:
[241] train's binary_logloss: 0.671442 valid's binary_logloss: 0.688118
bagging, val_score: 0.688118: 80%|######## | 8/10 [00:10<00:02, 1.31s/it][I 2020-09-27 04:54:46,218] Trial 34 finished with value: 0.6881176904104821 and parameters: {'bagging_fraction': 0.7616456433874508, 'bagging_freq': 4}. Best is trial 34 with value: 0.6881176904104821.
bagging, val_score: 0.688118: 80%|######## | 8/10 [00:10<00:02, 1.31s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009038 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.680934 valid's binary_logloss: 0.688963
[200] train's binary_logloss: 0.674092 valid's binary_logloss: 0.68854
[300] train's binary_logloss: 0.667767 valid's binary_logloss: 0.688887
Early stopping, best iteration is:
[208] train's binary_logloss: 0.673575 valid's binary_logloss: 0.688372
bagging, val_score: 0.688118: 90%|######### | 9/10 [00:11<00:01, 1.33s/it][I 2020-09-27 04:54:47,583] Trial 35 finished with value: 0.688371985997847 and parameters: {'bagging_fraction': 0.7538151922855485, 'bagging_freq': 4}. Best is trial 34 with value: 0.6881176904104821.
bagging, val_score: 0.688118: 90%|######### | 9/10 [00:11<00:01, 1.33s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001608 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68108 valid's binary_logloss: 0.689706
[200] train's binary_logloss: 0.674027 valid's binary_logloss: 0.689423
Early stopping, best iteration is:
[161] train's binary_logloss: 0.676626 valid's binary_logloss: 0.689208
bagging, val_score: 0.688118: 100%|##########| 10/10 [00:12<00:00, 1.29s/it][I 2020-09-27 04:54:48,782] Trial 36 finished with value: 0.6892084810261497 and parameters: {'bagging_fraction': 0.7387822613284701, 'bagging_freq': 4}. Best is trial 34 with value: 0.6881176904104821.
bagging, val_score: 0.688118: 100%|##########| 10/10 [00:12<00:00, 1.30s/it]
feature_fraction_stage2, val_score: 0.688118: 0%| | 0/6 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001509 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.680962 valid's binary_logloss: 0.689027
[200] train's binary_logloss: 0.673815 valid's binary_logloss: 0.689025
Early stopping, best iteration is:
[122] train's binary_logloss: 0.679212 valid's binary_logloss: 0.688791
feature_fraction_stage2, val_score: 0.688118: 17%|#6 | 1/6 [00:01<00:05, 1.04s/it][I 2020-09-27 04:54:49,838] Trial 37 finished with value: 0.6887910798116761 and parameters: {'feature_fraction': 0.88}. Best is trial 37 with value: 0.6887910798116761.
feature_fraction_stage2, val_score: 0.688118: 17%|#6 | 1/6 [00:01<00:05, 1.04s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008410 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681123 valid's binary_logloss: 0.688594
[200] train's binary_logloss: 0.674146 valid's binary_logloss: 0.688526
Early stopping, best iteration is:
[128] train's binary_logloss: 0.679071 valid's binary_logloss: 0.688123
feature_fraction_stage2, val_score: 0.688118: 33%|###3 | 2/6 [00:02<00:04, 1.03s/it][I 2020-09-27 04:54:50,833] Trial 38 finished with value: 0.6881231868561525 and parameters: {'feature_fraction': 0.7200000000000001}. Best is trial 38 with value: 0.6881231868561525.
feature_fraction_stage2, val_score: 0.688118: 33%|###3 | 2/6 [00:02<00:04, 1.03s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001586 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681076 valid's binary_logloss: 0.689153
[200] train's binary_logloss: 0.674066 valid's binary_logloss: 0.688436
[300] train's binary_logloss: 0.667548 valid's binary_logloss: 0.688896
Early stopping, best iteration is:
[241] train's binary_logloss: 0.671442 valid's binary_logloss: 0.688118
feature_fraction_stage2, val_score: 0.688118: 50%|##### | 3/6 [00:03<00:03, 1.18s/it][I 2020-09-27 04:54:52,366] Trial 39 finished with value: 0.6881176904104821 and parameters: {'feature_fraction': 0.8160000000000001}. Best is trial 39 with value: 0.6881176904104821.
feature_fraction_stage2, val_score: 0.688118: 50%|##### | 3/6 [00:03<00:03, 1.18s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011815 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681089 valid's binary_logloss: 0.689091
[200] train's binary_logloss: 0.674046 valid's binary_logloss: 0.68919
Early stopping, best iteration is:
[125] train's binary_logloss: 0.679183 valid's binary_logloss: 0.688872
feature_fraction_stage2, val_score: 0.688118: 67%|######6 | 4/6 [00:04<00:02, 1.13s/it][I 2020-09-27 04:54:53,387] Trial 40 finished with value: 0.6888723605544023 and parameters: {'feature_fraction': 0.784}. Best is trial 39 with value: 0.6881176904104821.
feature_fraction_stage2, val_score: 0.688118: 67%|######6 | 4/6 [00:04<00:02, 1.13s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008483 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681089 valid's binary_logloss: 0.689091
[200] train's binary_logloss: 0.674046 valid's binary_logloss: 0.68919
Early stopping, best iteration is:
[125] train's binary_logloss: 0.679183 valid's binary_logloss: 0.688872
feature_fraction_stage2, val_score: 0.688118: 83%|########3 | 5/6 [00:05<00:01, 1.10s/it][I 2020-09-27 04:54:54,428] Trial 41 finished with value: 0.6888723605544023 and parameters: {'feature_fraction': 0.7520000000000001}. Best is trial 39 with value: 0.6881176904104821.
feature_fraction_stage2, val_score: 0.688118: 83%|########3 | 5/6 [00:05<00:01, 1.10s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001564 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.680946 valid's binary_logloss: 0.689429
[200] train's binary_logloss: 0.67397 valid's binary_logloss: 0.689016
Early stopping, best iteration is:
[190] train's binary_logloss: 0.674674 valid's binary_logloss: 0.688999
feature_fraction_stage2, val_score: 0.688118: 100%|##########| 6/6 [00:06<00:00, 1.17s/it][I 2020-09-27 04:54:55,767] Trial 42 finished with value: 0.6889992541114389 and parameters: {'feature_fraction': 0.8480000000000001}. Best is trial 39 with value: 0.6881176904104821.
feature_fraction_stage2, val_score: 0.688118: 100%|##########| 6/6 [00:06<00:00, 1.16s/it]
regularization_factors, val_score: 0.688118: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009845 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681108 valid's binary_logloss: 0.689185
[200] train's binary_logloss: 0.674108 valid's binary_logloss: 0.68916
Early stopping, best iteration is:
[161] train's binary_logloss: 0.676632 valid's binary_logloss: 0.688793
regularization_factors, val_score: 0.688118: 5%|5 | 1/20 [00:01<00:22, 1.18s/it][I 2020-09-27 04:54:56,959] Trial 43 finished with value: 0.6887934098600209 and parameters: {'lambda_l1': 0.0384884910618084, 'lambda_l2': 1.204751116088014e-08}. Best is trial 43 with value: 0.6887934098600209.
regularization_factors, val_score: 0.688118: 5%|5 | 1/20 [00:01<00:22, 1.18s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001587 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68186 valid's binary_logloss: 0.689223
[200] train's binary_logloss: 0.675896 valid's binary_logloss: 0.689073
Early stopping, best iteration is:
[179] train's binary_logloss: 0.677134 valid's binary_logloss: 0.688878
regularization_factors, val_score: 0.688118: 10%|# | 2/20 [00:02<00:22, 1.22s/it][I 2020-09-27 04:54:58,283] Trial 44 finished with value: 0.6888779097684588 and parameters: {'lambda_l1': 1.6316340580802456e-08, 'lambda_l2': 9.98107290914818}. Best is trial 43 with value: 0.6887934098600209.
regularization_factors, val_score: 0.688118: 10%|# | 2/20 [00:02<00:22, 1.22s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008407 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681077 valid's binary_logloss: 0.689153
[200] train's binary_logloss: 0.674066 valid's binary_logloss: 0.688436
[300] train's binary_logloss: 0.667549 valid's binary_logloss: 0.688896
Early stopping, best iteration is:
[241] train's binary_logloss: 0.671443 valid's binary_logloss: 0.688118
regularization_factors, val_score: 0.688118: 15%|#5 | 3/20 [00:04<00:22, 1.31s/it][I 2020-09-27 04:54:59,812] Trial 45 finished with value: 0.688117656077511 and parameters: {'lambda_l1': 7.53282906565389e-08, 'lambda_l2': 0.000670356667923206}. Best is trial 45 with value: 0.688117656077511.
regularization_factors, val_score: 0.688118: 15%|#5 | 3/20 [00:04<00:22, 1.31s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001515 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681077 valid's binary_logloss: 0.689153
[200] train's binary_logloss: 0.674066 valid's binary_logloss: 0.688436
[300] train's binary_logloss: 0.667548 valid's binary_logloss: 0.688896
Early stopping, best iteration is:
[241] train's binary_logloss: 0.671443 valid's binary_logloss: 0.688118
regularization_factors, val_score: 0.688118: 20%|## | 4/20 [00:05<00:22, 1.39s/it][I 2020-09-27 04:55:01,389] Trial 46 finished with value: 0.6881176696895994 and parameters: {'lambda_l1': 2.804832879023663e-08, 'lambda_l2': 0.00040452126562955353}. Best is trial 45 with value: 0.688117656077511.
regularization_factors, val_score: 0.688118: 20%|## | 4/20 [00:05<00:22, 1.39s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008716 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681077 valid's binary_logloss: 0.689153
[200] train's binary_logloss: 0.674066 valid's binary_logloss: 0.688436
[300] train's binary_logloss: 0.667549 valid's binary_logloss: 0.688896
Early stopping, best iteration is:
[241] train's binary_logloss: 0.671443 valid's binary_logloss: 0.688118
regularization_factors, val_score: 0.688118: 25%|##5 | 5/20 [00:07<00:21, 1.43s/it][I 2020-09-27 04:55:02,896] Trial 47 finished with value: 0.6881176655585728 and parameters: {'lambda_l1': 1.5604063831480376e-08, 'lambda_l2': 0.0004850969258230225}. Best is trial 45 with value: 0.688117656077511.
regularization_factors, val_score: 0.688118: 25%|##5 | 5/20 [00:07<00:21, 1.43s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001586 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681077 valid's binary_logloss: 0.689153
[200] train's binary_logloss: 0.674066 valid's binary_logloss: 0.688436
[300] train's binary_logloss: 0.667548 valid's binary_logloss: 0.688896
Early stopping, best iteration is:
[241] train's binary_logloss: 0.671443 valid's binary_logloss: 0.688118
regularization_factors, val_score: 0.688118: 30%|### | 6/20 [00:08<00:20, 1.45s/it][I 2020-09-27 04:55:04,413] Trial 48 finished with value: 0.6881176673153139 and parameters: {'lambda_l1': 1.1804802009557963e-08, 'lambda_l2': 0.0004509060497221136}. Best is trial 45 with value: 0.688117656077511.
regularization_factors, val_score: 0.688118: 30%|### | 6/20 [00:08<00:20, 1.45s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.015933 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681077 valid's binary_logloss: 0.689153
[200] train's binary_logloss: 0.674066 valid's binary_logloss: 0.688436
[300] train's binary_logloss: 0.667549 valid's binary_logloss: 0.688896
Early stopping, best iteration is:
[241] train's binary_logloss: 0.671443 valid's binary_logloss: 0.688118
regularization_factors, val_score: 0.688118: 35%|###5 | 7/20 [00:10<00:19, 1.48s/it][I 2020-09-27 04:55:05,941] Trial 49 finished with value: 0.6881176626365213 and parameters: {'lambda_l1': 1.3636634898221364e-08, 'lambda_l2': 0.0005423517853601064}. Best is trial 45 with value: 0.688117656077511.
regularization_factors, val_score: 0.688118: 35%|###5 | 7/20 [00:10<00:19, 1.48s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010107 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681077 valid's binary_logloss: 0.689153
[200] train's binary_logloss: 0.674066 valid's binary_logloss: 0.688436
[300] train's binary_logloss: 0.667549 valid's binary_logloss: 0.688896
Early stopping, best iteration is:
[241] train's binary_logloss: 0.671443 valid's binary_logloss: 0.688118
regularization_factors, val_score: 0.688118: 40%|#### | 8/20 [00:11<00:17, 1.49s/it][I 2020-09-27 04:55:07,456] Trial 50 finished with value: 0.6881176473150824 and parameters: {'lambda_l1': 1.544452447154657e-08, 'lambda_l2': 0.000841525353074742}. Best is trial 50 with value: 0.6881176473150824.
regularization_factors, val_score: 0.688118: 40%|#### | 8/20 [00:11<00:17, 1.49s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009452 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681077 valid's binary_logloss: 0.689153
[200] train's binary_logloss: 0.674066 valid's binary_logloss: 0.688436
[300] train's binary_logloss: 0.667549 valid's binary_logloss: 0.688896
Early stopping, best iteration is:
[241] train's binary_logloss: 0.671443 valid's binary_logloss: 0.688118
regularization_factors, val_score: 0.688118: 45%|####5 | 9/20 [00:13<00:16, 1.50s/it][I 2020-09-27 04:55:08,986] Trial 51 finished with value: 0.6881176591565269 and parameters: {'lambda_l1': 1.1527007228477607e-08, 'lambda_l2': 0.0006103419407812519}. Best is trial 50 with value: 0.6881176473150824.
regularization_factors, val_score: 0.688118: 45%|####5 | 9/20 [00:13<00:16, 1.50s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001914 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681077 valid's binary_logloss: 0.689153
[200] train's binary_logloss: 0.674066 valid's binary_logloss: 0.688436
[300] train's binary_logloss: 0.667549 valid's binary_logloss: 0.688896
Early stopping, best iteration is:
[241] train's binary_logloss: 0.671443 valid's binary_logloss: 0.688118
regularization_factors, val_score: 0.688118: 50%|##### | 10/20 [00:14<00:15, 1.52s/it][I 2020-09-27 04:55:10,538] Trial 52 finished with value: 0.6881176497206014 and parameters: {'lambda_l1': 1.0151302543150935e-08, 'lambda_l2': 0.000794572846016082}. Best is trial 50 with value: 0.6881176473150824.
regularization_factors, val_score: 0.688118: 50%|##### | 10/20 [00:14<00:15, 1.52s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008731 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681077 valid's binary_logloss: 0.689153
[200] train's binary_logloss: 0.674066 valid's binary_logloss: 0.688436
[300] train's binary_logloss: 0.667549 valid's binary_logloss: 0.688896
Early stopping, best iteration is:
[241] train's binary_logloss: 0.671443 valid's binary_logloss: 0.688118
regularization_factors, val_score: 0.688118: 55%|#####5 | 11/20 [00:16<00:13, 1.52s/it][I 2020-09-27 04:55:12,058] Trial 53 finished with value: 0.6881176463901173 and parameters: {'lambda_l1': 1.6743100485936676e-08, 'lambda_l2': 0.0008594795917381767}. Best is trial 53 with value: 0.6881176463901173.
regularization_factors, val_score: 0.688118: 55%|#####5 | 11/20 [00:16<00:13, 1.52s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009990 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681077 valid's binary_logloss: 0.689152
[200] train's binary_logloss: 0.674067 valid's binary_logloss: 0.688436
[300] train's binary_logloss: 0.667549 valid's binary_logloss: 0.688896
Early stopping, best iteration is:
[241] train's binary_logloss: 0.671443 valid's binary_logloss: 0.688118
regularization_factors, val_score: 0.688118: 60%|###### | 12/20 [00:17<00:12, 1.52s/it][I 2020-09-27 04:55:13,591] Trial 54 finished with value: 0.6881176183828548 and parameters: {'lambda_l1': 1.3182804260599829e-08, 'lambda_l2': 0.0014065609790939724}. Best is trial 54 with value: 0.6881176183828548.
regularization_factors, val_score: 0.688118: 60%|###### | 12/20 [00:17<00:12, 1.52s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001677 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681004 valid's binary_logloss: 0.689046
[200] train's binary_logloss: 0.674082 valid's binary_logloss: 0.688629
[300] train's binary_logloss: 0.667629 valid's binary_logloss: 0.688981
Early stopping, best iteration is:
[208] train's binary_logloss: 0.673549 valid's binary_logloss: 0.68846
regularization_factors, val_score: 0.688118: 65%|######5 | 13/20 [00:19<00:10, 1.49s/it][I 2020-09-27 04:55:15,021] Trial 55 finished with value: 0.6884597615807989 and parameters: {'lambda_l1': 1.1894995962263343e-08, 'lambda_l2': 0.007778232962106016}. Best is trial 54 with value: 0.6881176183828548.
regularization_factors, val_score: 0.688118: 65%|######5 | 13/20 [00:19<00:10, 1.49s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001586 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681005 valid's binary_logloss: 0.689046
[200] train's binary_logloss: 0.674033 valid's binary_logloss: 0.688993
Early stopping, best iteration is:
[127] train's binary_logloss: 0.678994 valid's binary_logloss: 0.688676
regularization_factors, val_score: 0.688118: 70%|####### | 14/20 [00:20<00:08, 1.37s/it][I 2020-09-27 04:55:16,097] Trial 56 finished with value: 0.6886762297484497 and parameters: {'lambda_l1': 1.041076968285473e-06, 'lambda_l2': 0.011119399639455865}. Best is trial 54 with value: 0.6881176183828548.
regularization_factors, val_score: 0.688118: 70%|####### | 14/20 [00:20<00:08, 1.37s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001534 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681077 valid's binary_logloss: 0.689152
[200] train's binary_logloss: 0.67413 valid's binary_logloss: 0.68893
Early stopping, best iteration is:
[128] train's binary_logloss: 0.67903 valid's binary_logloss: 0.688711
regularization_factors, val_score: 0.688118: 75%|#######5 | 15/20 [00:21<00:06, 1.27s/it][I 2020-09-27 04:55:17,151] Trial 57 finished with value: 0.6887114735383899 and parameters: {'lambda_l1': 2.060141563353616e-07, 'lambda_l2': 0.004227573539355788}. Best is trial 54 with value: 0.6881176183828548.
regularization_factors, val_score: 0.688118: 75%|#######5 | 15/20 [00:21<00:06, 1.27s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001568 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681076 valid's binary_logloss: 0.689153
[200] train's binary_logloss: 0.674066 valid's binary_logloss: 0.688436
[300] train's binary_logloss: 0.667548 valid's binary_logloss: 0.688896
Early stopping, best iteration is:
[241] train's binary_logloss: 0.671442 valid's binary_logloss: 0.688118
regularization_factors, val_score: 0.688118: 80%|######## | 16/20 [00:22<00:05, 1.36s/it][I 2020-09-27 04:55:18,723] Trial 58 finished with value: 0.6881176889342293 and parameters: {'lambda_l1': 1.1056346896610525e-06, 'lambda_l2': 2.7165035646251157e-05}. Best is trial 54 with value: 0.6881176183828548.
regularization_factors, val_score: 0.688118: 80%|######## | 16/20 [00:22<00:05, 1.36s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001614 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681076 valid's binary_logloss: 0.689153
[200] train's binary_logloss: 0.674066 valid's binary_logloss: 0.688436
[300] train's binary_logloss: 0.667548 valid's binary_logloss: 0.688896
Early stopping, best iteration is:
[241] train's binary_logloss: 0.671442 valid's binary_logloss: 0.688118
regularization_factors, val_score: 0.688118: 85%|########5 | 17/20 [00:24<00:04, 1.42s/it][I 2020-09-27 04:55:20,289] Trial 59 finished with value: 0.6881176899381117 and parameters: {'lambda_l1': 1.4591391563330158e-07, 'lambda_l2': 9.037290952263432e-06}. Best is trial 54 with value: 0.6881176183828548.
regularization_factors, val_score: 0.688118: 85%|########5 | 17/20 [00:24<00:04, 1.42s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008880 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681099 valid's binary_logloss: 0.689028
[200] train's binary_logloss: 0.674171 valid's binary_logloss: 0.688687
[300] train's binary_logloss: 0.667693 valid's binary_logloss: 0.689247
Early stopping, best iteration is:
[208] train's binary_logloss: 0.67365 valid's binary_logloss: 0.688536
regularization_factors, val_score: 0.688118: 90%|######### | 18/20 [00:25<00:02, 1.43s/it][I 2020-09-27 04:55:21,723] Trial 60 finished with value: 0.6885355908433756 and parameters: {'lambda_l1': 1.3062472410110885e-08, 'lambda_l2': 0.007173981015512951}. Best is trial 54 with value: 0.6881176183828548.
regularization_factors, val_score: 0.688118: 90%|######### | 18/20 [00:25<00:02, 1.43s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001684 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681077 valid's binary_logloss: 0.689153
[200] train's binary_logloss: 0.674066 valid's binary_logloss: 0.688436
[300] train's binary_logloss: 0.667548 valid's binary_logloss: 0.688896
Early stopping, best iteration is:
[241] train's binary_logloss: 0.671443 valid's binary_logloss: 0.688118
regularization_factors, val_score: 0.688118: 95%|#########5| 19/20 [00:27<00:01, 1.47s/it][I 2020-09-27 04:55:23,279] Trial 61 finished with value: 0.6881176692604842 and parameters: {'lambda_l1': 1.0159772737706575e-08, 'lambda_l2': 0.0004128540970680604}. Best is trial 54 with value: 0.6881176183828548.
regularization_factors, val_score: 0.688118: 95%|#########5| 19/20 [00:27<00:01, 1.47s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001565 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681077 valid's binary_logloss: 0.689153
[200] train's binary_logloss: 0.674067 valid's binary_logloss: 0.688436
[300] train's binary_logloss: 0.667549 valid's binary_logloss: 0.688896
Early stopping, best iteration is:
[241] train's binary_logloss: 0.671443 valid's binary_logloss: 0.688118
regularization_factors, val_score: 0.688118: 100%|##########| 20/20 [00:29<00:00, 1.48s/it][I 2020-09-27 04:55:24,804] Trial 62 finished with value: 0.6881176378261832 and parameters: {'lambda_l1': 1.2256608662872115e-08, 'lambda_l2': 0.0010266987268648666}. Best is trial 54 with value: 0.6881176183828548.
regularization_factors, val_score: 0.688118: 100%|##########| 20/20 [00:29<00:00, 1.45s/it]
min_data_in_leaf, val_score: 0.688118: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001754 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681082 valid's binary_logloss: 0.68904
[200] train's binary_logloss: 0.674109 valid's binary_logloss: 0.689221
Early stopping, best iteration is:
[127] train's binary_logloss: 0.679064 valid's binary_logloss: 0.688843
min_data_in_leaf, val_score: 0.688118: 20%|## | 1/5 [00:01<00:04, 1.11s/it][I 2020-09-27 04:55:25,927] Trial 63 finished with value: 0.6888427522145505 and parameters: {'min_child_samples': 25}. Best is trial 63 with value: 0.6888427522145505.
min_data_in_leaf, val_score: 0.688118: 20%|## | 1/5 [00:01<00:04, 1.11s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001638 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68099 valid's binary_logloss: 0.689568
[200] train's binary_logloss: 0.67383 valid's binary_logloss: 0.68945
Early stopping, best iteration is:
[139] train's binary_logloss: 0.678037 valid's binary_logloss: 0.689261
min_data_in_leaf, val_score: 0.688118: 40%|#### | 2/5 [00:02<00:03, 1.16s/it][I 2020-09-27 04:55:27,218] Trial 64 finished with value: 0.6892610898091048 and parameters: {'min_child_samples': 5}. Best is trial 63 with value: 0.6888427522145505.
min_data_in_leaf, val_score: 0.688118: 40%|#### | 2/5 [00:02<00:03, 1.16s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002000 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681202 valid's binary_logloss: 0.689481
[200] train's binary_logloss: 0.674521 valid's binary_logloss: 0.689204
Early stopping, best iteration is:
[133] train's binary_logloss: 0.678939 valid's binary_logloss: 0.688898
min_data_in_leaf, val_score: 0.688118: 60%|###### | 3/5 [00:03<00:02, 1.18s/it][I 2020-09-27 04:55:28,425] Trial 65 finished with value: 0.688897822488096 and parameters: {'min_child_samples': 50}. Best is trial 63 with value: 0.6888427522145505.
min_data_in_leaf, val_score: 0.688118: 60%|###### | 3/5 [00:03<00:02, 1.18s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001549 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681041 valid's binary_logloss: 0.68912
[200] train's binary_logloss: 0.673932 valid's binary_logloss: 0.689165
[300] train's binary_logloss: 0.667472 valid's binary_logloss: 0.689403
Early stopping, best iteration is:
[223] train's binary_logloss: 0.672471 valid's binary_logloss: 0.688731
min_data_in_leaf, val_score: 0.688118: 80%|######## | 4/5 [00:05<00:01, 1.28s/it][I 2020-09-27 04:55:29,951] Trial 66 finished with value: 0.6887314309367388 and parameters: {'min_child_samples': 10}. Best is trial 66 with value: 0.6887314309367388.
min_data_in_leaf, val_score: 0.688118: 80%|######## | 4/5 [00:05<00:01, 1.28s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001500 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407
[LightGBM] [Info] Start training from score 0.006407
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.681418 valid's binary_logloss: 0.688989
[200] train's binary_logloss: 0.675177 valid's binary_logloss: 0.688794
[300] train's binary_logloss: 0.669107 valid's binary_logloss: 0.688977
Early stopping, best iteration is:
[242] train's binary_logloss: 0.672665 valid's binary_logloss: 0.688301
min_data_in_leaf, val_score: 0.688118: 100%|##########| 5/5 [00:06<00:00, 1.38s/it][I 2020-09-27 04:55:31,562] Trial 67 finished with value: 0.6883012831237107 and parameters: {'min_child_samples': 100}. Best is trial 67 with value: 0.6883012831237107.
min_data_in_leaf, val_score: 0.688118: 100%|##########| 5/5 [00:06<00:00, 1.35s/it]
Fold : 4
[I 2020-09-27 04:55:31,677] A new study created in memory with name: no-name-9f8068ee-5ed7-40e8-b31b-fe76cdedea06
feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008972 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664047 valid's binary_logloss: 0.690512
Early stopping, best iteration is:
[55] train's binary_logloss: 0.674377 valid's binary_logloss: 0.689789
feature_fraction, val_score: 0.689789: 14%|#4 | 1/7 [00:00<00:05, 1.05it/s][I 2020-09-27 04:55:32,636] Trial 0 finished with value: 0.6897889878374779 and parameters: {'feature_fraction': 0.6}. Best is trial 0 with value: 0.6897889878374779.
feature_fraction, val_score: 0.689789: 14%|#4 | 1/7 [00:00<00:05, 1.05it/s][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000871 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.666186 valid's binary_logloss: 0.690817
Early stopping, best iteration is:
[41] train's binary_logloss: 0.679056 valid's binary_logloss: 0.69044
feature_fraction, val_score: 0.689789: 29%|##8 | 2/7 [00:01<00:04, 1.14it/s][I 2020-09-27 04:55:33,334] Trial 1 finished with value: 0.6904400471655459 and parameters: {'feature_fraction': 0.4}. Best is trial 0 with value: 0.6897889878374779.
feature_fraction, val_score: 0.689789: 29%|##8 | 2/7 [00:01<00:04, 1.14it/s][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000896 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.66495 valid's binary_logloss: 0.68987
Early stopping, best iteration is:
[42] train's binary_logloss: 0.678016 valid's binary_logloss: 0.689555
feature_fraction, val_score: 0.689555: 43%|####2 | 3/7 [00:02<00:03, 1.19it/s][I 2020-09-27 04:55:34,092] Trial 2 finished with value: 0.6895553636305107 and parameters: {'feature_fraction': 0.5}. Best is trial 2 with value: 0.6895553636305107.
feature_fraction, val_score: 0.689555: 43%|####2 | 3/7 [00:02<00:03, 1.19it/s][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.014383 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663818 valid's binary_logloss: 0.690283
Early stopping, best iteration is:
[60] train's binary_logloss: 0.672948 valid's binary_logloss: 0.689614
feature_fraction, val_score: 0.689555: 57%|#####7 | 4/7 [00:03<00:02, 1.20it/s][I 2020-09-27 04:55:34,910] Trial 3 finished with value: 0.689613792253809 and parameters: {'feature_fraction': 0.7}. Best is trial 2 with value: 0.6895553636305107.
feature_fraction, val_score: 0.689555: 57%|#####7 | 4/7 [00:03<00:02, 1.20it/s][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.007348 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663516 valid's binary_logloss: 0.691084
Early stopping, best iteration is:
[56] train's binary_logloss: 0.673734 valid's binary_logloss: 0.690379
feature_fraction, val_score: 0.689555: 71%|#######1 | 5/7 [00:04<00:01, 1.18it/s][I 2020-09-27 04:55:35,802] Trial 4 finished with value: 0.6903789869460358 and parameters: {'feature_fraction': 0.8}. Best is trial 2 with value: 0.6895553636305107.
feature_fraction, val_score: 0.689555: 71%|#######1 | 5/7 [00:04<00:01, 1.18it/s][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001795 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.662299 valid's binary_logloss: 0.689516
Early stopping, best iteration is:
[43] train's binary_logloss: 0.676528 valid's binary_logloss: 0.689224
feature_fraction, val_score: 0.689224: 86%|########5 | 6/7 [00:05<00:00, 1.15it/s][I 2020-09-27 04:55:36,710] Trial 5 finished with value: 0.689223941248766 and parameters: {'feature_fraction': 1.0}. Best is trial 5 with value: 0.689223941248766.
feature_fraction, val_score: 0.689224: 86%|########5 | 6/7 [00:05<00:00, 1.15it/s][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.005320 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663036 valid's binary_logloss: 0.690663
Early stopping, best iteration is:
[45] train's binary_logloss: 0.676494 valid's binary_logloss: 0.689393
feature_fraction, val_score: 0.689224: 100%|##########| 7/7 [00:06<00:00, 1.10it/s][I 2020-09-27 04:55:37,710] Trial 6 finished with value: 0.6893925766913008 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 5 with value: 0.689223941248766.
feature_fraction, val_score: 0.689224: 100%|##########| 7/7 [00:06<00:00, 1.16it/s]
num_leaves, val_score: 0.689224: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001743 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.55132 valid's binary_logloss: 0.697443
Early stopping, best iteration is:
[11] train's binary_logloss: 0.668724 valid's binary_logloss: 0.691123
num_leaves, val_score: 0.689224: 5%|5 | 1/20 [00:01<00:26, 1.38s/it][I 2020-09-27 04:55:39,105] Trial 7 finished with value: 0.6911225737618327 and parameters: {'num_leaves': 200}. Best is trial 7 with value: 0.6911225737618327.
num_leaves, val_score: 0.689224: 5%|5 | 1/20 [00:01<00:26, 1.38s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001580 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.571031 valid's binary_logloss: 0.698616
Early stopping, best iteration is:
[13] train's binary_logloss: 0.669076 valid's binary_logloss: 0.690967
num_leaves, val_score: 0.689224: 10%|# | 2/20 [00:02<00:25, 1.39s/it][I 2020-09-27 04:55:40,522] Trial 8 finished with value: 0.690966972517542 and parameters: {'num_leaves': 165}. Best is trial 8 with value: 0.690966972517542.
num_leaves, val_score: 0.689224: 10%|# | 2/20 [00:02<00:25, 1.39s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002611 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.608797 valid's binary_logloss: 0.692752
Early stopping, best iteration is:
[19] train's binary_logloss: 0.670075 valid's binary_logloss: 0.690291
num_leaves, val_score: 0.689224: 15%|#5 | 3/20 [00:04<00:23, 1.36s/it][I 2020-09-27 04:55:41,828] Trial 9 finished with value: 0.690291129092192 and parameters: {'num_leaves': 104}. Best is trial 9 with value: 0.690291129092192.
num_leaves, val_score: 0.689224: 15%|#5 | 3/20 [00:04<00:23, 1.36s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001684 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687085 valid's binary_logloss: 0.689639
[200] train's binary_logloss: 0.684519 valid's binary_logloss: 0.689387
[300] train's binary_logloss: 0.682492 valid's binary_logloss: 0.689351
Early stopping, best iteration is:
[285] train's binary_logloss: 0.682778 valid's binary_logloss: 0.689263
num_leaves, val_score: 0.689224: 20%|## | 4/20 [00:05<00:23, 1.47s/it][I 2020-09-27 04:55:43,548] Trial 10 finished with value: 0.6892629508973852 and parameters: {'num_leaves': 4}. Best is trial 10 with value: 0.6892629508973852.
num_leaves, val_score: 0.689224: 20%|## | 4/20 [00:05<00:23, 1.47s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002867 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687085 valid's binary_logloss: 0.689639
[200] train's binary_logloss: 0.684519 valid's binary_logloss: 0.689387
[300] train's binary_logloss: 0.682492 valid's binary_logloss: 0.689351
Early stopping, best iteration is:
[285] train's binary_logloss: 0.682778 valid's binary_logloss: 0.689263
num_leaves, val_score: 0.689224: 25%|##5 | 5/20 [00:07<00:24, 1.62s/it][I 2020-09-27 04:55:45,500] Trial 11 finished with value: 0.6892629508973852 and parameters: {'num_leaves': 4}. Best is trial 10 with value: 0.6892629508973852.
num_leaves, val_score: 0.689224: 25%|##5 | 5/20 [00:07<00:24, 1.62s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.003844 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.670769 valid's binary_logloss: 0.690136
Early stopping, best iteration is:
[57] train's binary_logloss: 0.67811 valid's binary_logloss: 0.689439
num_leaves, val_score: 0.689224: 30%|### | 6/20 [00:08<00:20, 1.46s/it][I 2020-09-27 04:55:46,606] Trial 12 finished with value: 0.6894394773112021 and parameters: {'num_leaves': 21}. Best is trial 10 with value: 0.6892629508973852.
num_leaves, val_score: 0.689224: 30%|### | 6/20 [00:08<00:20, 1.46s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002745 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.632184 valid's binary_logloss: 0.690411
Early stopping, best iteration is:
[41] train's binary_logloss: 0.662296 valid's binary_logloss: 0.689656
num_leaves, val_score: 0.689224: 35%|###5 | 7/20 [00:09<00:17, 1.35s/it][I 2020-09-27 04:55:47,703] Trial 13 finished with value: 0.6896563634195803 and parameters: {'num_leaves': 71}. Best is trial 10 with value: 0.6892629508973852.
num_leaves, val_score: 0.689224: 35%|###5 | 7/20 [00:09<00:17, 1.35s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001681 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.528217 valid's binary_logloss: 0.698531
Early stopping, best iteration is:
[10] train's binary_logloss: 0.667084 valid's binary_logloss: 0.692043
num_leaves, val_score: 0.689224: 40%|#### | 8/20 [00:11<00:16, 1.39s/it][I 2020-09-27 04:55:49,169] Trial 14 finished with value: 0.6920431771973239 and parameters: {'num_leaves': 242}. Best is trial 10 with value: 0.6892629508973852.
num_leaves, val_score: 0.689224: 40%|#### | 8/20 [00:11<00:16, 1.39s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001737 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.647395 valid's binary_logloss: 0.692015
Early stopping, best iteration is:
[18] train's binary_logloss: 0.680717 valid's binary_logloss: 0.689923
num_leaves, val_score: 0.689224: 45%|####5 | 9/20 [00:12<00:13, 1.19s/it][I 2020-09-27 04:55:49,908] Trial 15 finished with value: 0.6899230440367558 and parameters: {'num_leaves': 50}. Best is trial 10 with value: 0.6892629508973852.
num_leaves, val_score: 0.689224: 45%|####5 | 9/20 [00:12<00:13, 1.19s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001534 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.599344 valid's binary_logloss: 0.693711
Early stopping, best iteration is:
[20] train's binary_logloss: 0.666348 valid's binary_logloss: 0.689963
num_leaves, val_score: 0.689224: 50%|##### | 10/20 [00:13<00:11, 1.16s/it][I 2020-09-27 04:55:50,983] Trial 16 finished with value: 0.6899633518471027 and parameters: {'num_leaves': 119}. Best is trial 10 with value: 0.6892629508973852.
num_leaves, val_score: 0.689224: 50%|##### | 10/20 [00:13<00:11, 1.16s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001595 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688413 valid's binary_logloss: 0.689959
[200] train's binary_logloss: 0.686555 valid's binary_logloss: 0.689381
[300] train's binary_logloss: 0.685253 valid's binary_logloss: 0.688961
[400] train's binary_logloss: 0.684142 valid's binary_logloss: 0.689082
Early stopping, best iteration is:
[301] train's binary_logloss: 0.685239 valid's binary_logloss: 0.688951
num_leaves, val_score: 0.688951: 55%|#####5 | 11/20 [00:14<00:11, 1.24s/it][I 2020-09-27 04:55:52,401] Trial 17 finished with value: 0.6889511525087615 and parameters: {'num_leaves': 3}. Best is trial 17 with value: 0.6889511525087615.
num_leaves, val_score: 0.688951: 55%|#####5 | 11/20 [00:14<00:11, 1.24s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002030 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.571031 valid's binary_logloss: 0.698616
Early stopping, best iteration is:
[13] train's binary_logloss: 0.669076 valid's binary_logloss: 0.690967
num_leaves, val_score: 0.688951: 60%|###### | 12/20 [00:16<00:10, 1.27s/it][I 2020-09-27 04:55:53,737] Trial 18 finished with value: 0.690966972517542 and parameters: {'num_leaves': 165}. Best is trial 17 with value: 0.6889511525087615.
num_leaves, val_score: 0.688951: 60%|###### | 12/20 [00:16<00:10, 1.27s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001718 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.628315 valid's binary_logloss: 0.692025
Early stopping, best iteration is:
[25] train's binary_logloss: 0.670941 valid's binary_logloss: 0.689909
num_leaves, val_score: 0.688951: 65%|######5 | 13/20 [00:16<00:08, 1.17s/it][I 2020-09-27 04:55:54,692] Trial 19 finished with value: 0.6899089284634762 and parameters: {'num_leaves': 76}. Best is trial 17 with value: 0.6889511525087615.
num_leaves, val_score: 0.688951: 65%|######5 | 13/20 [00:16<00:08, 1.17s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001740 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656752 valid's binary_logloss: 0.690382
Early stopping, best iteration is:
[52] train's binary_logloss: 0.670809 valid's binary_logloss: 0.689428
num_leaves, val_score: 0.688951: 70%|####### | 14/20 [00:17<00:06, 1.08s/it][I 2020-09-27 04:55:55,563] Trial 20 finished with value: 0.6894277899501294 and parameters: {'num_leaves': 38}. Best is trial 17 with value: 0.6889511525087615.
num_leaves, val_score: 0.688951: 70%|####### | 14/20 [00:17<00:06, 1.08s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001722 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687085 valid's binary_logloss: 0.689639
[200] train's binary_logloss: 0.684519 valid's binary_logloss: 0.689387
[300] train's binary_logloss: 0.682492 valid's binary_logloss: 0.689351
Early stopping, best iteration is:
[285] train's binary_logloss: 0.682778 valid's binary_logloss: 0.689263
num_leaves, val_score: 0.688951: 75%|#######5 | 15/20 [00:19<00:05, 1.17s/it][I 2020-09-27 04:55:56,939] Trial 21 finished with value: 0.6892629508973853 and parameters: {'num_leaves': 4}. Best is trial 17 with value: 0.6889511525087615.
num_leaves, val_score: 0.688951: 75%|#######5 | 15/20 [00:19<00:05, 1.17s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.005007 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687085 valid's binary_logloss: 0.689639
[200] train's binary_logloss: 0.684519 valid's binary_logloss: 0.689387
[300] train's binary_logloss: 0.682492 valid's binary_logloss: 0.689351
Early stopping, best iteration is:
[285] train's binary_logloss: 0.682778 valid's binary_logloss: 0.689263
num_leaves, val_score: 0.688951: 80%|######## | 16/20 [00:20<00:04, 1.25s/it][I 2020-09-27 04:55:58,373] Trial 22 finished with value: 0.6892629508973852 and parameters: {'num_leaves': 4}. Best is trial 17 with value: 0.6889511525087615.
num_leaves, val_score: 0.688951: 80%|######## | 16/20 [00:20<00:04, 1.25s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001737 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.634712 valid's binary_logloss: 0.690552
Early stopping, best iteration is:
[28] train's binary_logloss: 0.671201 valid's binary_logloss: 0.689554
num_leaves, val_score: 0.688951: 85%|########5 | 17/20 [00:21<00:03, 1.17s/it][I 2020-09-27 04:55:59,359] Trial 23 finished with value: 0.6895543062404332 and parameters: {'num_leaves': 67}. Best is trial 17 with value: 0.6889511525087615.
num_leaves, val_score: 0.688951: 85%|########5 | 17/20 [00:21<00:03, 1.17s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001626 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.660898 valid's binary_logloss: 0.690623
Early stopping, best iteration is:
[44] train's binary_logloss: 0.675571 valid's binary_logloss: 0.689922
num_leaves, val_score: 0.688951: 90%|######### | 18/20 [00:22<00:02, 1.09s/it][I 2020-09-27 04:56:00,277] Trial 24 finished with value: 0.6899216528961626 and parameters: {'num_leaves': 33}. Best is trial 17 with value: 0.6889511525087615.
num_leaves, val_score: 0.688951: 90%|######### | 18/20 [00:22<00:02, 1.09s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002059 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.614335 valid's binary_logloss: 0.691319
Early stopping, best iteration is:
[19] train's binary_logloss: 0.67155 valid's binary_logloss: 0.689698
num_leaves, val_score: 0.688951: 95%|#########5| 19/20 [00:23<00:01, 1.08s/it][I 2020-09-27 04:56:01,311] Trial 25 finished with value: 0.6896983317380696 and parameters: {'num_leaves': 96}. Best is trial 17 with value: 0.6889511525087615.
num_leaves, val_score: 0.688951: 95%|#########5| 19/20 [00:23<00:01, 1.08s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001688 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.58378 valid's binary_logloss: 0.69401
Early stopping, best iteration is:
[20] train's binary_logloss: 0.662024 valid's binary_logloss: 0.690131
num_leaves, val_score: 0.688951: 100%|##########| 20/20 [00:24<00:00, 1.11s/it][I 2020-09-27 04:56:02,500] Trial 26 finished with value: 0.6901310985346112 and parameters: {'num_leaves': 143}. Best is trial 17 with value: 0.6889511525087615.
num_leaves, val_score: 0.688951: 100%|##########| 20/20 [00:24<00:00, 1.24s/it]
bagging, val_score: 0.688951: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013453 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688279 valid's binary_logloss: 0.689921
[200] train's binary_logloss: 0.686342 valid's binary_logloss: 0.68936
[300] train's binary_logloss: 0.685002 valid's binary_logloss: 0.688756
[400] train's binary_logloss: 0.683806 valid's binary_logloss: 0.688689
Early stopping, best iteration is:
[358] train's binary_logloss: 0.684279 valid's binary_logloss: 0.688644
bagging, val_score: 0.688644: 10%|# | 1/10 [00:01<00:16, 1.81s/it][I 2020-09-27 04:56:04,317] Trial 27 finished with value: 0.6886437067357495 and parameters: {'bagging_fraction': 0.8690258816156133, 'bagging_freq': 3}. Best is trial 27 with value: 0.6886437067357495.
bagging, val_score: 0.688644: 10%|# | 1/10 [00:01<00:16, 1.81s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001668 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688299 valid's binary_logloss: 0.689981
[200] train's binary_logloss: 0.686384 valid's binary_logloss: 0.689407
[300] train's binary_logloss: 0.684946 valid's binary_logloss: 0.689175
Early stopping, best iteration is:
[276] train's binary_logloss: 0.685288 valid's binary_logloss: 0.689089
bagging, val_score: 0.688644: 20%|## | 2/10 [00:03<00:13, 1.73s/it][I 2020-09-27 04:56:05,876] Trial 28 finished with value: 0.6890889360892993 and parameters: {'bagging_fraction': 0.8933236439302734, 'bagging_freq': 3}. Best is trial 27 with value: 0.6886437067357495.
bagging, val_score: 0.688644: 20%|## | 2/10 [00:03<00:13, 1.73s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001756 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688343 valid's binary_logloss: 0.689852
[200] train's binary_logloss: 0.686411 valid's binary_logloss: 0.689334
[300] train's binary_logloss: 0.685065 valid's binary_logloss: 0.688979
[400] train's binary_logloss: 0.683898 valid's binary_logloss: 0.689086
Early stopping, best iteration is:
[315] train's binary_logloss: 0.684866 valid's binary_logloss: 0.688904
bagging, val_score: 0.688644: 30%|### | 3/10 [00:05<00:12, 1.73s/it][I 2020-09-27 04:56:07,585] Trial 29 finished with value: 0.6889043328094836 and parameters: {'bagging_fraction': 0.9252835167638693, 'bagging_freq': 3}. Best is trial 27 with value: 0.6886437067357495.
bagging, val_score: 0.688644: 30%|### | 3/10 [00:05<00:12, 1.73s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001792 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688307 valid's binary_logloss: 0.689977
[200] train's binary_logloss: 0.686382 valid's binary_logloss: 0.689334
[300] train's binary_logloss: 0.684994 valid's binary_logloss: 0.689157
[400] train's binary_logloss: 0.683796 valid's binary_logloss: 0.688976
Early stopping, best iteration is:
[399] train's binary_logloss: 0.683807 valid's binary_logloss: 0.688973
bagging, val_score: 0.688644: 40%|#### | 4/10 [00:07<00:10, 1.82s/it][I 2020-09-27 04:56:09,625] Trial 30 finished with value: 0.688972596287182 and parameters: {'bagging_fraction': 0.9070779201672944, 'bagging_freq': 3}. Best is trial 27 with value: 0.6886437067357495.
bagging, val_score: 0.688644: 40%|#### | 4/10 [00:07<00:10, 1.82s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001680 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688296 valid's binary_logloss: 0.689913
[200] train's binary_logloss: 0.686352 valid's binary_logloss: 0.689474
[300] train's binary_logloss: 0.684941 valid's binary_logloss: 0.689008
[400] train's binary_logloss: 0.683731 valid's binary_logloss: 0.689001
Early stopping, best iteration is:
[361] train's binary_logloss: 0.684198 valid's binary_logloss: 0.688907
bagging, val_score: 0.688644: 50%|##### | 5/10 [00:08<00:09, 1.84s/it][I 2020-09-27 04:56:11,504] Trial 31 finished with value: 0.6889071822847319 and parameters: {'bagging_fraction': 0.9123574981524005, 'bagging_freq': 3}. Best is trial 27 with value: 0.6886437067357495.
bagging, val_score: 0.688644: 50%|##### | 5/10 [00:08<00:09, 1.84s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002215 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688308 valid's binary_logloss: 0.689935
[200] train's binary_logloss: 0.686414 valid's binary_logloss: 0.689366
[300] train's binary_logloss: 0.685039 valid's binary_logloss: 0.689039
[400] train's binary_logloss: 0.683854 valid's binary_logloss: 0.689061
Early stopping, best iteration is:
[313] train's binary_logloss: 0.684864 valid's binary_logloss: 0.688989
bagging, val_score: 0.688644: 60%|###### | 6/10 [00:10<00:07, 1.80s/it][I 2020-09-27 04:56:13,223] Trial 32 finished with value: 0.6889888493170525 and parameters: {'bagging_fraction': 0.9247906381445511, 'bagging_freq': 3}. Best is trial 27 with value: 0.6886437067357495.
bagging, val_score: 0.688644: 60%|###### | 6/10 [00:10<00:07, 1.80s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001684 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613
[200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995
[300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612
[400] train's binary_logloss: 0.683551 valid's binary_logloss: 0.688705
Early stopping, best iteration is:
[330] train's binary_logloss: 0.684394 valid's binary_logloss: 0.68851
bagging, val_score: 0.688510: 70%|####### | 7/10 [00:13<00:06, 2.23s/it][I 2020-09-27 04:56:16,468] Trial 33 finished with value: 0.6885102663693616 and parameters: {'bagging_fraction': 0.7258338472094388, 'bagging_freq': 1}. Best is trial 33 with value: 0.6885102663693616.
bagging, val_score: 0.688510: 70%|####### | 7/10 [00:13<00:06, 2.23s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.003707 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687986 valid's binary_logloss: 0.689762
[200] train's binary_logloss: 0.686029 valid's binary_logloss: 0.689311
[300] train's binary_logloss: 0.684604 valid's binary_logloss: 0.689125
Early stopping, best iteration is:
[243] train's binary_logloss: 0.685398 valid's binary_logloss: 0.689001
bagging, val_score: 0.688510: 80%|######## | 8/10 [00:15<00:04, 2.04s/it][I 2020-09-27 04:56:18,063] Trial 34 finished with value: 0.6890013155830245 and parameters: {'bagging_fraction': 0.5748859423991363, 'bagging_freq': 1}. Best is trial 33 with value: 0.6885102663693616.
bagging, val_score: 0.688510: 80%|######## | 8/10 [00:15<00:04, 2.04s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.003161 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688221 valid's binary_logloss: 0.689816
[200] train's binary_logloss: 0.686335 valid's binary_logloss: 0.688896
Early stopping, best iteration is:
[196] train's binary_logloss: 0.686393 valid's binary_logloss: 0.688862
bagging, val_score: 0.688510: 90%|######### | 9/10 [00:16<00:01, 1.81s/it][I 2020-09-27 04:56:19,330] Trial 35 finished with value: 0.6888615132015019 and parameters: {'bagging_fraction': 0.7649026913055617, 'bagging_freq': 6}. Best is trial 33 with value: 0.6885102663693616.
bagging, val_score: 0.688510: 90%|######### | 9/10 [00:16<00:01, 1.81s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001740 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688211 valid's binary_logloss: 0.689704
[200] train's binary_logloss: 0.686297 valid's binary_logloss: 0.689225
[300] train's binary_logloss: 0.684963 valid's binary_logloss: 0.688871
Early stopping, best iteration is:
[229] train's binary_logloss: 0.685871 valid's binary_logloss: 0.688848
bagging, val_score: 0.688510: 100%|##########| 10/10 [00:18<00:00, 1.68s/it][I 2020-09-27 04:56:20,711] Trial 36 finished with value: 0.6888483568488261 and parameters: {'bagging_fraction': 0.7251434202313678, 'bagging_freq': 7}. Best is trial 33 with value: 0.6885102663693616.
bagging, val_score: 0.688510: 100%|##########| 10/10 [00:18<00:00, 1.82s/it]
feature_fraction_stage2, val_score: 0.688510: 0%| | 0/3 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001750 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613
[200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995
[300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612
[400] train's binary_logloss: 0.683551 valid's binary_logloss: 0.688705
Early stopping, best iteration is:
[330] train's binary_logloss: 0.684394 valid's binary_logloss: 0.68851
feature_fraction_stage2, val_score: 0.688510: 33%|###3 | 1/3 [00:01<00:03, 1.59s/it][I 2020-09-27 04:56:22,315] Trial 37 finished with value: 0.6885102663693616 and parameters: {'feature_fraction': 0.9840000000000001}. Best is trial 37 with value: 0.6885102663693616.
feature_fraction_stage2, val_score: 0.688510: 33%|###3 | 1/3 [00:01<00:03, 1.59s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001619 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688132 valid's binary_logloss: 0.689669
[200] train's binary_logloss: 0.686204 valid's binary_logloss: 0.689145
[300] train's binary_logloss: 0.684787 valid's binary_logloss: 0.688653
[400] train's binary_logloss: 0.683619 valid's binary_logloss: 0.688636
Early stopping, best iteration is:
[392] train's binary_logloss: 0.683706 valid's binary_logloss: 0.688545
feature_fraction_stage2, val_score: 0.688510: 67%|######6 | 2/3 [00:03<00:01, 1.66s/it][I 2020-09-27 04:56:24,121] Trial 38 finished with value: 0.6885449109441385 and parameters: {'feature_fraction': 0.92}. Best is trial 37 with value: 0.6885102663693616.
feature_fraction_stage2, val_score: 0.688510: 67%|######6 | 2/3 [00:03<00:01, 1.66s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001677 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68813 valid's binary_logloss: 0.689528
[200] train's binary_logloss: 0.686156 valid's binary_logloss: 0.688931
[300] train's binary_logloss: 0.684723 valid's binary_logloss: 0.688714
[400] train's binary_logloss: 0.683522 valid's binary_logloss: 0.688698
Early stopping, best iteration is:
[382] train's binary_logloss: 0.683729 valid's binary_logloss: 0.688541
feature_fraction_stage2, val_score: 0.688510: 100%|##########| 3/3 [00:05<00:00, 1.69s/it][I 2020-09-27 04:56:25,892] Trial 39 finished with value: 0.688541416001024 and parameters: {'feature_fraction': 0.9520000000000001}. Best is trial 37 with value: 0.6885102663693616.
feature_fraction_stage2, val_score: 0.688510: 100%|##########| 3/3 [00:05<00:00, 1.72s/it]
regularization_factors, val_score: 0.688510: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.005767 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613
[200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995
[300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612
[400] train's binary_logloss: 0.683573 valid's binary_logloss: 0.68868
Early stopping, best iteration is:
[380] train's binary_logloss: 0.683811 valid's binary_logloss: 0.688528
regularization_factors, val_score: 0.688510: 5%|5 | 1/20 [00:01<00:33, 1.76s/it][I 2020-09-27 04:56:27,666] Trial 40 finished with value: 0.6885276922964685 and parameters: {'lambda_l1': 1.5626678097623024e-08, 'lambda_l2': 0.0035654183554903016}. Best is trial 40 with value: 0.6885276922964685.
regularization_factors, val_score: 0.688510: 5%|5 | 1/20 [00:01<00:33, 1.76s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002094 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613
[200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995
[300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612
[400] train's binary_logloss: 0.683572 valid's binary_logloss: 0.68868
Early stopping, best iteration is:
[380] train's binary_logloss: 0.683811 valid's binary_logloss: 0.688528
regularization_factors, val_score: 0.688510: 10%|# | 2/20 [00:03<00:31, 1.76s/it][I 2020-09-27 04:56:29,430] Trial 41 finished with value: 0.6885276861626745 and parameters: {'lambda_l1': 3.870259482083587e-08, 'lambda_l2': 0.0031399198971774163}. Best is trial 41 with value: 0.6885276861626745.
regularization_factors, val_score: 0.688510: 10%|# | 2/20 [00:03<00:31, 1.76s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001738 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613
[200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995
[300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612
[400] train's binary_logloss: 0.683572 valid's binary_logloss: 0.68868
Early stopping, best iteration is:
[380] train's binary_logloss: 0.683811 valid's binary_logloss: 0.688528
regularization_factors, val_score: 0.688510: 15%|#5 | 3/20 [00:05<00:29, 1.76s/it][I 2020-09-27 04:56:31,201] Trial 42 finished with value: 0.6885276847893826 and parameters: {'lambda_l1': 1.2562118956422349e-08, 'lambda_l2': 0.0030450124632060223}. Best is trial 42 with value: 0.6885276847893826.
regularization_factors, val_score: 0.688510: 15%|#5 | 3/20 [00:05<00:29, 1.76s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013796 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613
[200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995
[300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612
[400] train's binary_logloss: 0.683573 valid's binary_logloss: 0.68868
Early stopping, best iteration is:
[380] train's binary_logloss: 0.683811 valid's binary_logloss: 0.688528
regularization_factors, val_score: 0.688510: 20%|## | 4/20 [00:07<00:28, 1.81s/it][I 2020-09-27 04:56:33,111] Trial 43 finished with value: 0.6885276913121847 and parameters: {'lambda_l1': 1.1600996445683937e-08, 'lambda_l2': 0.0034970196571133183}. Best is trial 42 with value: 0.6885276847893826.
regularization_factors, val_score: 0.688510: 20%|## | 4/20 [00:07<00:28, 1.81s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009630 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613
[200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995
[300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612
[400] train's binary_logloss: 0.683572 valid's binary_logloss: 0.68868
Early stopping, best iteration is:
[380] train's binary_logloss: 0.683811 valid's binary_logloss: 0.688528
regularization_factors, val_score: 0.688510: 25%|##5 | 5/20 [00:09<00:27, 1.84s/it][I 2020-09-27 04:56:35,011] Trial 44 finished with value: 0.6885276777308208 and parameters: {'lambda_l1': 1.5284605069192858e-08, 'lambda_l2': 0.002555456660087826}. Best is trial 44 with value: 0.6885276777308208.
regularization_factors, val_score: 0.688510: 25%|##5 | 5/20 [00:09<00:27, 1.84s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001670 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613
[200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995
[300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612
[400] train's binary_logloss: 0.683572 valid's binary_logloss: 0.68868
Early stopping, best iteration is:
[380] train's binary_logloss: 0.683811 valid's binary_logloss: 0.688528
regularization_factors, val_score: 0.688510: 30%|### | 6/20 [00:10<00:25, 1.81s/it][I 2020-09-27 04:56:36,753] Trial 45 finished with value: 0.6885276789659555 and parameters: {'lambda_l1': 1.761891864374289e-08, 'lambda_l2': 0.0026412614459624425}. Best is trial 44 with value: 0.6885276777308208.
regularization_factors, val_score: 0.688510: 30%|### | 6/20 [00:10<00:25, 1.81s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010747 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688144 valid's binary_logloss: 0.689555
[200] train's binary_logloss: 0.686223 valid's binary_logloss: 0.688989
[300] train's binary_logloss: 0.684835 valid's binary_logloss: 0.688807
[400] train's binary_logloss: 0.683655 valid's binary_logloss: 0.688801
Early stopping, best iteration is:
[366] train's binary_logloss: 0.684032 valid's binary_logloss: 0.688577
regularization_factors, val_score: 0.688510: 35%|###5 | 7/20 [00:12<00:23, 1.82s/it][I 2020-09-27 04:56:38,598] Trial 46 finished with value: 0.6885772867402516 and parameters: {'lambda_l1': 0.06434172773582825, 'lambda_l2': 4.068285530948775e-06}. Best is trial 44 with value: 0.6885276777308208.
regularization_factors, val_score: 0.688510: 35%|###5 | 7/20 [00:12<00:23, 1.82s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001651 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688154 valid's binary_logloss: 0.689765
[200] train's binary_logloss: 0.686207 valid's binary_logloss: 0.689123
[300] train's binary_logloss: 0.684829 valid's binary_logloss: 0.688815
[400] train's binary_logloss: 0.683653 valid's binary_logloss: 0.688858
Early stopping, best iteration is:
[378] train's binary_logloss: 0.68391 valid's binary_logloss: 0.688721
regularization_factors, val_score: 0.688510: 40%|#### | 8/20 [00:14<00:21, 1.81s/it][I 2020-09-27 04:56:40,403] Trial 47 finished with value: 0.6887213435654922 and parameters: {'lambda_l1': 1.2218484580158208e-06, 'lambda_l2': 0.5657158290313663}. Best is trial 44 with value: 0.6885276777308208.
regularization_factors, val_score: 0.688510: 40%|#### | 8/20 [00:14<00:21, 1.81s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001718 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613
[200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995
[300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612
[400] train's binary_logloss: 0.683551 valid's binary_logloss: 0.688705
Early stopping, best iteration is:
[330] train's binary_logloss: 0.684394 valid's binary_logloss: 0.68851
regularization_factors, val_score: 0.688510: 45%|####5 | 9/20 [00:16<00:19, 1.77s/it][I 2020-09-27 04:56:42,076] Trial 48 finished with value: 0.6885102681025668 and parameters: {'lambda_l1': 1.1710448780595842e-05, 'lambda_l2': 6.970955109724959e-05}. Best is trial 48 with value: 0.6885102681025668.
regularization_factors, val_score: 0.688510: 45%|####5 | 9/20 [00:16<00:19, 1.77s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001926 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613
[200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995
[300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612
[400] train's binary_logloss: 0.683551 valid's binary_logloss: 0.688705
Early stopping, best iteration is:
[330] train's binary_logloss: 0.684394 valid's binary_logloss: 0.68851
regularization_factors, val_score: 0.688510: 50%|##### | 10/20 [00:17<00:17, 1.71s/it][I 2020-09-27 04:56:43,655] Trial 49 finished with value: 0.6885102671718292 and parameters: {'lambda_l1': 2.6398830760982305e-05, 'lambda_l2': 8.347668509260632e-06}. Best is trial 49 with value: 0.6885102671718292.
regularization_factors, val_score: 0.688510: 50%|##### | 10/20 [00:17<00:17, 1.71s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001747 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613
[200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995
[300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612
[400] train's binary_logloss: 0.683572 valid's binary_logloss: 0.68868
Early stopping, best iteration is:
[380] train's binary_logloss: 0.683811 valid's binary_logloss: 0.688528
regularization_factors, val_score: 0.688510: 55%|#####5 | 11/20 [00:19<00:15, 1.73s/it][I 2020-09-27 04:56:45,435] Trial 50 finished with value: 0.6885276427800883 and parameters: {'lambda_l1': 9.524089932248691e-05, 'lambda_l2': 2.4636927718157655e-06}. Best is trial 49 with value: 0.6885102671718292.
regularization_factors, val_score: 0.688510: 55%|#####5 | 11/20 [00:19<00:15, 1.73s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001941 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613
[200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995
[300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612
[400] train's binary_logloss: 0.683572 valid's binary_logloss: 0.68868
Early stopping, best iteration is:
[380] train's binary_logloss: 0.683811 valid's binary_logloss: 0.688528
regularization_factors, val_score: 0.688510: 60%|###### | 12/20 [00:21<00:13, 1.75s/it][I 2020-09-27 04:56:47,207] Trial 51 finished with value: 0.6885276426289637 and parameters: {'lambda_l1': 8.799095559564636e-05, 'lambda_l2': 2.1427847942855786e-06}. Best is trial 49 with value: 0.6885102671718292.
regularization_factors, val_score: 0.688510: 60%|###### | 12/20 [00:21<00:13, 1.75s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001695 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613
[200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995
[300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612
[400] train's binary_logloss: 0.683572 valid's binary_logloss: 0.68868
Early stopping, best iteration is:
[380] train's binary_logloss: 0.683811 valid's binary_logloss: 0.688528
regularization_factors, val_score: 0.688510: 65%|######5 | 13/20 [00:23<00:12, 1.75s/it][I 2020-09-27 04:56:48,982] Trial 52 finished with value: 0.688527642704103 and parameters: {'lambda_l1': 9.263556836103949e-05, 'lambda_l2': 8.942562462981873e-07}. Best is trial 49 with value: 0.6885102671718292.
regularization_factors, val_score: 0.688510: 65%|######5 | 13/20 [00:23<00:12, 1.75s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001690 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613
[200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995
[300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612
[400] train's binary_logloss: 0.683572 valid's binary_logloss: 0.68868
Early stopping, best iteration is:
[380] train's binary_logloss: 0.683811 valid's binary_logloss: 0.688528
regularization_factors, val_score: 0.688510: 70%|####### | 14/20 [00:24<00:10, 1.75s/it][I 2020-09-27 04:56:50,729] Trial 53 finished with value: 0.6885276427926035 and parameters: {'lambda_l1': 9.755819246036098e-05, 'lambda_l2': 1.883877450611394e-08}. Best is trial 49 with value: 0.6885102671718292.
regularization_factors, val_score: 0.688510: 70%|####### | 14/20 [00:24<00:10, 1.75s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011495 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613
[200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995
[300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612
[400] train's binary_logloss: 0.683551 valid's binary_logloss: 0.688705
Early stopping, best iteration is:
[330] train's binary_logloss: 0.684394 valid's binary_logloss: 0.68851
regularization_factors, val_score: 0.688510: 75%|#######5 | 15/20 [00:26<00:08, 1.74s/it][I 2020-09-27 04:56:52,434] Trial 54 finished with value: 0.6885102665832229 and parameters: {'lambda_l1': 7.204721356400207e-06, 'lambda_l2': 1.9515753697920597e-06}. Best is trial 54 with value: 0.6885102665832229.
regularization_factors, val_score: 0.688510: 75%|#######5 | 15/20 [00:26<00:08, 1.74s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001867 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613
[200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995
[300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612
[400] train's binary_logloss: 0.683551 valid's binary_logloss: 0.688705
Early stopping, best iteration is:
[330] train's binary_logloss: 0.684394 valid's binary_logloss: 0.68851
regularization_factors, val_score: 0.688510: 80%|######## | 16/20 [00:28<00:06, 1.69s/it][I 2020-09-27 04:56:54,005] Trial 55 finished with value: 0.6885102672275221 and parameters: {'lambda_l1': 2.7150467223950745e-06, 'lambda_l2': 3.804591403735543e-05}. Best is trial 54 with value: 0.6885102665832229.
regularization_factors, val_score: 0.688510: 80%|######## | 16/20 [00:28<00:06, 1.69s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001819 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613
[200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995
[300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612
[400] train's binary_logloss: 0.683551 valid's binary_logloss: 0.688705
Early stopping, best iteration is:
[330] train's binary_logloss: 0.684394 valid's binary_logloss: 0.68851
regularization_factors, val_score: 0.688510: 85%|########5 | 17/20 [00:29<00:04, 1.67s/it][I 2020-09-27 04:56:55,619] Trial 56 finished with value: 0.6885102669786968 and parameters: {'lambda_l1': 1.4913904132624564e-06, 'lambda_l2': 2.7565544406824723e-05}. Best is trial 54 with value: 0.6885102665832229.
regularization_factors, val_score: 0.688510: 85%|########5 | 17/20 [00:29<00:04, 1.67s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001735 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613
[200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995
[300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612
[400] train's binary_logloss: 0.683551 valid's binary_logloss: 0.688705
Early stopping, best iteration is:
[330] train's binary_logloss: 0.684394 valid's binary_logloss: 0.68851
regularization_factors, val_score: 0.688510: 90%|######### | 18/20 [00:31<00:03, 1.63s/it][I 2020-09-27 04:56:57,171] Trial 57 finished with value: 0.6885102674416176 and parameters: {'lambda_l1': 7.533020720507337e-07, 'lambda_l2': 5.04819339961594e-05}. Best is trial 54 with value: 0.6885102665832229.
regularization_factors, val_score: 0.688510: 90%|######### | 18/20 [00:31<00:03, 1.63s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001739 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613
[200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995
[300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612
[400] train's binary_logloss: 0.683551 valid's binary_logloss: 0.688705
Early stopping, best iteration is:
[330] train's binary_logloss: 0.684394 valid's binary_logloss: 0.68851
regularization_factors, val_score: 0.688510: 95%|#########5| 19/20 [00:32<00:01, 1.63s/it][I 2020-09-27 04:56:58,787] Trial 58 finished with value: 0.6885102664101546 and parameters: {'lambda_l1': 1.6489132998257337e-06, 'lambda_l2': 7.130405244958287e-08}. Best is trial 58 with value: 0.6885102664101546.
regularization_factors, val_score: 0.688510: 95%|#########5| 19/20 [00:32<00:01, 1.63s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001734 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613
[200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995
[300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688613
[400] train's binary_logloss: 0.683573 valid's binary_logloss: 0.68868
Early stopping, best iteration is:
[380] train's binary_logloss: 0.683811 valid's binary_logloss: 0.688528
regularization_factors, val_score: 0.688510: 100%|##########| 20/20 [00:34<00:00, 1.67s/it][I 2020-09-27 04:57:00,567] Trial 59 finished with value: 0.6885278039688245 and parameters: {'lambda_l1': 0.008341325209129346, 'lambda_l2': 6.068578447774457e-08}. Best is trial 58 with value: 0.6885102664101546.
regularization_factors, val_score: 0.688510: 100%|##########| 20/20 [00:34<00:00, 1.73s/it]
min_data_in_leaf, val_score: 0.688510: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002224 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613
[200] train's binary_logloss: 0.686175 valid's binary_logloss: 0.68904
[300] train's binary_logloss: 0.684779 valid's binary_logloss: 0.68875
[400] train's binary_logloss: 0.683563 valid's binary_logloss: 0.68872
[500] train's binary_logloss: 0.682433 valid's binary_logloss: 0.688523
[600] train's binary_logloss: 0.681362 valid's binary_logloss: 0.688598
Early stopping, best iteration is:
[540] train's binary_logloss: 0.682021 valid's binary_logloss: 0.688417
min_data_in_leaf, val_score: 0.688417: 20%|## | 1/5 [00:02<00:09, 2.34s/it][I 2020-09-27 04:57:02,917] Trial 60 finished with value: 0.6884174980111903 and parameters: {'min_child_samples': 25}. Best is trial 60 with value: 0.6884174980111903.
min_data_in_leaf, val_score: 0.688417: 20%|## | 1/5 [00:02<00:09, 2.34s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001678 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613
[200] train's binary_logloss: 0.686186 valid's binary_logloss: 0.68902
[300] train's binary_logloss: 0.684775 valid's binary_logloss: 0.688899
[400] train's binary_logloss: 0.683559 valid's binary_logloss: 0.688863
Early stopping, best iteration is:
[378] train's binary_logloss: 0.68381 valid's binary_logloss: 0.68872
min_data_in_leaf, val_score: 0.688417: 40%|#### | 2/5 [00:04<00:06, 2.16s/it][I 2020-09-27 04:57:04,665] Trial 61 finished with value: 0.6887198054735056 and parameters: {'min_child_samples': 50}. Best is trial 60 with value: 0.6884174980111903.
min_data_in_leaf, val_score: 0.688417: 40%|#### | 2/5 [00:04<00:06, 2.16s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001665 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688189 valid's binary_logloss: 0.689923
[200] train's binary_logloss: 0.686316 valid's binary_logloss: 0.689183
[300] train's binary_logloss: 0.684943 valid's binary_logloss: 0.689041
Early stopping, best iteration is:
[254] train's binary_logloss: 0.685519 valid's binary_logloss: 0.688957
min_data_in_leaf, val_score: 0.688417: 60%|###### | 3/5 [00:05<00:03, 1.91s/it][I 2020-09-27 04:57:05,984] Trial 62 finished with value: 0.6889570125436887 and parameters: {'min_child_samples': 100}. Best is trial 60 with value: 0.6884174980111903.
min_data_in_leaf, val_score: 0.688417: 60%|###### | 3/5 [00:05<00:03, 1.91s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001746 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613
[200] train's binary_logloss: 0.686178 valid's binary_logloss: 0.688982
[300] train's binary_logloss: 0.684774 valid's binary_logloss: 0.688717
[400] train's binary_logloss: 0.683552 valid's binary_logloss: 0.688696
Early stopping, best iteration is:
[383] train's binary_logloss: 0.683767 valid's binary_logloss: 0.688615
min_data_in_leaf, val_score: 0.688417: 80%|######## | 4/5 [00:07<00:01, 1.86s/it][I 2020-09-27 04:57:07,748] Trial 63 finished with value: 0.6886149362033982 and parameters: {'min_child_samples': 10}. Best is trial 60 with value: 0.6884174980111903.
min_data_in_leaf, val_score: 0.688417: 80%|######## | 4/5 [00:07<00:01, 1.86s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001731 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4688
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493
[LightGBM] [Info] Start training from score 0.006493
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613
[200] train's binary_logloss: 0.686183 valid's binary_logloss: 0.688973
[300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688751
[400] train's binary_logloss: 0.683531 valid's binary_logloss: 0.688863
Early stopping, best iteration is:
[306] train's binary_logloss: 0.684696 valid's binary_logloss: 0.688722
min_data_in_leaf, val_score: 0.688417: 100%|##########| 5/5 [00:08<00:00, 1.76s/it][I 2020-09-27 04:57:09,249] Trial 64 finished with value: 0.6887218892158072 and parameters: {'min_child_samples': 5}. Best is trial 60 with value: 0.6884174980111903.
min_data_in_leaf, val_score: 0.688417: 100%|##########| 5/5 [00:08<00:00, 1.73s/it]
Fold : 5
[I 2020-09-27 04:57:09,427] A new study created in memory with name: no-name-1b7c624a-f0b2-4aea-8fc8-aa6993df7768
feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001589 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.662759 valid's binary_logloss: 0.691799
Early stopping, best iteration is:
[36] train's binary_logloss: 0.678598 valid's binary_logloss: 0.69056
feature_fraction, val_score: 0.690560: 14%|#4 | 1/7 [00:00<00:05, 1.06it/s][I 2020-09-27 04:57:10,381] Trial 0 finished with value: 0.6905601497910684 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 0 with value: 0.6905601497910684.
feature_fraction, val_score: 0.690560: 14%|#4 | 1/7 [00:00<00:05, 1.06it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.007342 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.66369 valid's binary_logloss: 0.691565
Early stopping, best iteration is:
[25] train's binary_logloss: 0.68226 valid's binary_logloss: 0.690559
feature_fraction, val_score: 0.690559: 29%|##8 | 2/7 [00:01<00:04, 1.15it/s][I 2020-09-27 04:57:11,086] Trial 1 finished with value: 0.6905590100900298 and parameters: {'feature_fraction': 0.7}. Best is trial 1 with value: 0.6905590100900298.
feature_fraction, val_score: 0.690559: 29%|##8 | 2/7 [00:01<00:04, 1.15it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011035 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664057 valid's binary_logloss: 0.690869
Early stopping, best iteration is:
[40] train's binary_logloss: 0.678145 valid's binary_logloss: 0.690182
feature_fraction, val_score: 0.690182: 43%|####2 | 3/7 [00:02<00:03, 1.23it/s][I 2020-09-27 04:57:11,765] Trial 2 finished with value: 0.6901820889991936 and parameters: {'feature_fraction': 0.6}. Best is trial 2 with value: 0.6901820889991936.
feature_fraction, val_score: 0.690182: 43%|####2 | 3/7 [00:02<00:03, 1.23it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001692 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664852 valid's binary_logloss: 0.691416
Early stopping, best iteration is:
[35] train's binary_logloss: 0.679773 valid's binary_logloss: 0.690316
feature_fraction, val_score: 0.690182: 57%|#####7 | 4/7 [00:03<00:02, 1.29it/s][I 2020-09-27 04:57:12,459] Trial 3 finished with value: 0.690316333441943 and parameters: {'feature_fraction': 0.5}. Best is trial 2 with value: 0.6901820889991936.
feature_fraction, val_score: 0.690182: 57%|#####7 | 4/7 [00:03<00:02, 1.29it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.006609 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663021 valid's binary_logloss: 0.690361
Early stopping, best iteration is:
[48] train's binary_logloss: 0.675428 valid's binary_logloss: 0.689984
feature_fraction, val_score: 0.689984: 71%|#######1 | 5/7 [00:03<00:01, 1.27it/s][I 2020-09-27 04:57:13,274] Trial 4 finished with value: 0.6899844980770151 and parameters: {'feature_fraction': 0.8}. Best is trial 4 with value: 0.6899844980770151.
feature_fraction, val_score: 0.689984: 71%|#######1 | 5/7 [00:03<00:01, 1.27it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001693 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.662006 valid's binary_logloss: 0.691157
Early stopping, best iteration is:
[40] train's binary_logloss: 0.67696 valid's binary_logloss: 0.690385
feature_fraction, val_score: 0.689984: 86%|########5 | 6/7 [00:04<00:00, 1.26it/s][I 2020-09-27 04:57:14,069] Trial 5 finished with value: 0.690384793694187 and parameters: {'feature_fraction': 1.0}. Best is trial 4 with value: 0.6899844980770151.
feature_fraction, val_score: 0.689984: 86%|########5 | 6/7 [00:04<00:00, 1.26it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000701 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.666045 valid's binary_logloss: 0.691233
Early stopping, best iteration is:
[38] train's binary_logloss: 0.67965 valid's binary_logloss: 0.690765
feature_fraction, val_score: 0.689984: 100%|##########| 7/7 [00:05<00:00, 1.31it/s][I 2020-09-27 04:57:14,764] Trial 6 finished with value: 0.6907650282143057 and parameters: {'feature_fraction': 0.4}. Best is trial 4 with value: 0.6899844980770151.
feature_fraction, val_score: 0.689984: 100%|##########| 7/7 [00:05<00:00, 1.31it/s]
num_leaves, val_score: 0.689984: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008471 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.532446 valid's binary_logloss: 0.701071
Early stopping, best iteration is:
[12] train's binary_logloss: 0.663644 valid's binary_logloss: 0.692101
num_leaves, val_score: 0.689984: 5%|5 | 1/20 [00:01<00:24, 1.28s/it][I 2020-09-27 04:57:16,058] Trial 7 finished with value: 0.6921010909265272 and parameters: {'num_leaves': 246}. Best is trial 7 with value: 0.6921010909265272.
num_leaves, val_score: 0.689984: 5%|5 | 1/20 [00:01<00:24, 1.28s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011999 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.541655 valid's binary_logloss: 0.698738
Early stopping, best iteration is:
[19] train's binary_logloss: 0.651466 valid's binary_logloss: 0.691604
num_leaves, val_score: 0.689984: 10%|# | 2/20 [00:02<00:23, 1.29s/it][I 2020-09-27 04:57:17,368] Trial 8 finished with value: 0.6916040591247711 and parameters: {'num_leaves': 228}. Best is trial 8 with value: 0.6916040591247711.
num_leaves, val_score: 0.689984: 10%|# | 2/20 [00:02<00:23, 1.29s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001462 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.528502 valid's binary_logloss: 0.699513
Early stopping, best iteration is:
[12] train's binary_logloss: 0.662669 valid's binary_logloss: 0.69248
num_leaves, val_score: 0.689984: 15%|#5 | 3/20 [00:04<00:22, 1.34s/it][I 2020-09-27 04:57:18,834] Trial 9 finished with value: 0.6924803382332924 and parameters: {'num_leaves': 254}. Best is trial 8 with value: 0.6916040591247711.
num_leaves, val_score: 0.689984: 15%|#5 | 3/20 [00:04<00:22, 1.34s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001476 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.683804 valid's binary_logloss: 0.690293
[200] train's binary_logloss: 0.67902 valid's binary_logloss: 0.690554
Early stopping, best iteration is:
[108] train's binary_logloss: 0.683362 valid's binary_logloss: 0.690199
num_leaves, val_score: 0.689984: 20%|## | 4/20 [00:04<00:19, 1.19s/it][I 2020-09-27 04:57:19,664] Trial 10 finished with value: 0.6901988057794263 and parameters: {'num_leaves': 7}. Best is trial 10 with value: 0.6901988057794263.
num_leaves, val_score: 0.689984: 20%|## | 4/20 [00:04<00:19, 1.19s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001573 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.677338 valid's binary_logloss: 0.690393
[200] train's binary_logloss: 0.667531 valid's binary_logloss: 0.69092
Early stopping, best iteration is:
[138] train's binary_logloss: 0.673519 valid's binary_logloss: 0.690353
num_leaves, val_score: 0.689984: 25%|##5 | 5/20 [00:05<00:17, 1.15s/it][I 2020-09-27 04:57:20,726] Trial 11 finished with value: 0.6903529337197906 and parameters: {'num_leaves': 14}. Best is trial 10 with value: 0.6901988057794263.
num_leaves, val_score: 0.689984: 25%|##5 | 5/20 [00:05<00:17, 1.15s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001618 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.607789 valid's binary_logloss: 0.695
Early stopping, best iteration is:
[16] train's binary_logloss: 0.673018 valid's binary_logloss: 0.691131
num_leaves, val_score: 0.689984: 30%|### | 6/20 [00:06<00:15, 1.12s/it][I 2020-09-27 04:57:21,759] Trial 12 finished with value: 0.6911309288725374 and parameters: {'num_leaves': 110}. Best is trial 10 with value: 0.6901988057794263.
num_leaves, val_score: 0.689984: 30%|### | 6/20 [00:06<00:15, 1.12s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002136 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.584924 valid's binary_logloss: 0.695787
Early stopping, best iteration is:
[13] train's binary_logloss: 0.672012 valid's binary_logloss: 0.691214
num_leaves, val_score: 0.689984: 35%|###5 | 7/20 [00:08<00:14, 1.12s/it][I 2020-09-27 04:57:22,903] Trial 13 finished with value: 0.6912141405715722 and parameters: {'num_leaves': 148}. Best is trial 10 with value: 0.6901988057794263.
num_leaves, val_score: 0.689984: 35%|###5 | 7/20 [00:08<00:14, 1.12s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001517 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.632471 valid's binary_logloss: 0.693122
Early stopping, best iteration is:
[14] train's binary_logloss: 0.680218 valid's binary_logloss: 0.691475
num_leaves, val_score: 0.689984: 40%|#### | 8/20 [00:08<00:12, 1.03s/it][I 2020-09-27 04:57:23,723] Trial 14 finished with value: 0.6914751164482753 and parameters: {'num_leaves': 72}. Best is trial 10 with value: 0.6901988057794263.
num_leaves, val_score: 0.689984: 40%|#### | 8/20 [00:08<00:12, 1.03s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001436 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.563691 valid's binary_logloss: 0.697937
Early stopping, best iteration is:
[9] train's binary_logloss: 0.67486 valid's binary_logloss: 0.691909
num_leaves, val_score: 0.689984: 45%|####5 | 9/20 [00:10<00:11, 1.09s/it][I 2020-09-27 04:57:24,931] Trial 15 finished with value: 0.6919085281880217 and parameters: {'num_leaves': 185}. Best is trial 10 with value: 0.6901988057794263.
num_leaves, val_score: 0.689984: 45%|####5 | 9/20 [00:10<00:11, 1.09s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001412 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.658128 valid's binary_logloss: 0.690955
Early stopping, best iteration is:
[30] train's binary_logloss: 0.678602 valid's binary_logloss: 0.690303
num_leaves, val_score: 0.689984: 50%|##### | 10/20 [00:10<00:09, 1.02it/s][I 2020-09-27 04:57:25,662] Trial 16 finished with value: 0.6903025101431474 and parameters: {'num_leaves': 37}. Best is trial 10 with value: 0.6901988057794263.
num_leaves, val_score: 0.689984: 50%|##### | 10/20 [00:10<00:09, 1.02it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011796 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.562179 valid's binary_logloss: 0.697416
Early stopping, best iteration is:
[13] train's binary_logloss: 0.667864 valid's binary_logloss: 0.691752
num_leaves, val_score: 0.689984: 55%|#####5 | 11/20 [00:12<00:09, 1.04s/it][I 2020-09-27 04:57:26,832] Trial 17 finished with value: 0.6917519955746995 and parameters: {'num_leaves': 187}. Best is trial 10 with value: 0.6901988057794263.
num_leaves, val_score: 0.689984: 55%|#####5 | 11/20 [00:12<00:09, 1.04s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002032 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.614951 valid's binary_logloss: 0.693854
Early stopping, best iteration is:
[26] train's binary_logloss: 0.665369 valid's binary_logloss: 0.690866
num_leaves, val_score: 0.689984: 60%|###### | 12/20 [00:13<00:08, 1.04s/it][I 2020-09-27 04:57:27,890] Trial 18 finished with value: 0.6908659763700707 and parameters: {'num_leaves': 99}. Best is trial 10 with value: 0.6901988057794263.
num_leaves, val_score: 0.689984: 60%|###### | 12/20 [00:13<00:08, 1.04s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001505 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.637367 valid's binary_logloss: 0.693163
Early stopping, best iteration is:
[32] train's binary_logloss: 0.669343 valid's binary_logloss: 0.690841
num_leaves, val_score: 0.689984: 65%|######5 | 13/20 [00:14<00:06, 1.00it/s][I 2020-09-27 04:57:28,782] Trial 19 finished with value: 0.6908408834731546 and parameters: {'num_leaves': 66}. Best is trial 10 with value: 0.6901988057794263.
num_leaves, val_score: 0.689984: 65%|######5 | 13/20 [00:14<00:06, 1.00it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008072 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.575818 valid's binary_logloss: 0.696901
Early stopping, best iteration is:
[13] train's binary_logloss: 0.669822 valid's binary_logloss: 0.691657
num_leaves, val_score: 0.689984: 70%|####### | 14/20 [00:15<00:06, 1.02s/it][I 2020-09-27 04:57:29,853] Trial 20 finished with value: 0.6916567124710575 and parameters: {'num_leaves': 166}. Best is trial 10 with value: 0.6901988057794263.
num_leaves, val_score: 0.689984: 70%|####### | 14/20 [00:15<00:06, 1.02s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008152 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687062 valid's binary_logloss: 0.690297
[200] train's binary_logloss: 0.684539 valid's binary_logloss: 0.690125
Early stopping, best iteration is:
[195] train's binary_logloss: 0.684633 valid's binary_logloss: 0.690069
num_leaves, val_score: 0.689984: 75%|#######5 | 15/20 [00:16<00:05, 1.04s/it][I 2020-09-27 04:57:30,950] Trial 21 finished with value: 0.6900688079787631 and parameters: {'num_leaves': 4}. Best is trial 21 with value: 0.6900688079787631.
num_leaves, val_score: 0.689984: 75%|#######5 | 15/20 [00:16<00:05, 1.04s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002005 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.660627 valid's binary_logloss: 0.690947
Early stopping, best iteration is:
[34] train's binary_logloss: 0.678295 valid's binary_logloss: 0.690073
num_leaves, val_score: 0.689984: 80%|######## | 16/20 [00:16<00:03, 1.04it/s][I 2020-09-27 04:57:31,722] Trial 22 finished with value: 0.6900727900021961 and parameters: {'num_leaves': 34}. Best is trial 21 with value: 0.6900688079787631.
num_leaves, val_score: 0.689984: 80%|######## | 16/20 [00:16<00:03, 1.04it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001577 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.653581 valid's binary_logloss: 0.692192
Early stopping, best iteration is:
[26] train's binary_logloss: 0.67856 valid's binary_logloss: 0.690524
num_leaves, val_score: 0.689984: 85%|########5 | 17/20 [00:17<00:02, 1.10it/s][I 2020-09-27 04:57:32,506] Trial 23 finished with value: 0.6905242489383265 and parameters: {'num_leaves': 43}. Best is trial 21 with value: 0.6900688079787631.
num_leaves, val_score: 0.689984: 85%|########5 | 17/20 [00:17<00:02, 1.10it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001600 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.682861 valid's binary_logloss: 0.690265
Early stopping, best iteration is:
[76] train's binary_logloss: 0.684408 valid's binary_logloss: 0.690209
num_leaves, val_score: 0.689984: 90%|######### | 18/20 [00:18<00:01, 1.16it/s][I 2020-09-27 04:57:33,262] Trial 24 finished with value: 0.6902091508032887 and parameters: {'num_leaves': 8}. Best is trial 21 with value: 0.6900688079787631.
num_leaves, val_score: 0.689984: 90%|######### | 18/20 [00:18<00:01, 1.16it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.014414 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.656074 valid's binary_logloss: 0.691055
Early stopping, best iteration is:
[34] train's binary_logloss: 0.676266 valid's binary_logloss: 0.690265
num_leaves, val_score: 0.689984: 95%|#########5| 19/20 [00:19<00:00, 1.19it/s][I 2020-09-27 04:57:34,040] Trial 25 finished with value: 0.6902649750375632 and parameters: {'num_leaves': 40}. Best is trial 21 with value: 0.6900688079787631.
num_leaves, val_score: 0.689984: 95%|#########5| 19/20 [00:19<00:00, 1.19it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001563 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.630456 valid's binary_logloss: 0.693
Early stopping, best iteration is:
[21] train's binary_logloss: 0.67434 valid's binary_logloss: 0.691243
num_leaves, val_score: 0.689984: 100%|##########| 20/20 [00:20<00:00, 1.17it/s][I 2020-09-27 04:57:34,925] Trial 26 finished with value: 0.6912431590178266 and parameters: {'num_leaves': 75}. Best is trial 21 with value: 0.6900688079787631.
num_leaves, val_score: 0.689984: 100%|##########| 20/20 [00:20<00:00, 1.01s/it]
bagging, val_score: 0.689984: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001562 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.662713 valid's binary_logloss: 0.690614
Early stopping, best iteration is:
[61] train's binary_logloss: 0.672055 valid's binary_logloss: 0.690011
bagging, val_score: 0.689984: 10%|# | 1/10 [00:01<00:09, 1.06s/it][I 2020-09-27 04:57:35,998] Trial 27 finished with value: 0.6900113794489978 and parameters: {'bagging_fraction': 0.7744675089831714, 'bagging_freq': 3}. Best is trial 27 with value: 0.6900113794489978.
bagging, val_score: 0.689984: 10%|# | 1/10 [00:01<00:09, 1.06s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001550 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663033 valid's binary_logloss: 0.691575
Early stopping, best iteration is:
[65] train's binary_logloss: 0.671129 valid's binary_logloss: 0.690781
bagging, val_score: 0.689984: 20%|## | 2/10 [00:02<00:08, 1.10s/it][I 2020-09-27 04:57:37,192] Trial 28 finished with value: 0.6907811511387683 and parameters: {'bagging_fraction': 0.7835892578642625, 'bagging_freq': 3}. Best is trial 27 with value: 0.6900113794489978.
bagging, val_score: 0.689984: 20%|## | 2/10 [00:02<00:08, 1.10s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009896 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664095 valid's binary_logloss: 0.692608
Early stopping, best iteration is:
[35] train's binary_logloss: 0.67955 valid's binary_logloss: 0.690369
bagging, val_score: 0.689984: 30%|### | 3/10 [00:03<00:07, 1.05s/it][I 2020-09-27 04:57:38,119] Trial 29 finished with value: 0.6903685689943968 and parameters: {'bagging_fraction': 0.4766066067413073, 'bagging_freq': 1}. Best is trial 27 with value: 0.6900113794489978.
bagging, val_score: 0.689984: 30%|### | 3/10 [00:03<00:07, 1.05s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008136 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.662586 valid's binary_logloss: 0.691009
Early stopping, best iteration is:
[26] train's binary_logloss: 0.68167 valid's binary_logloss: 0.690305
bagging, val_score: 0.689984: 40%|#### | 4/10 [00:03<00:05, 1.02it/s][I 2020-09-27 04:57:38,929] Trial 30 finished with value: 0.6903053053602763 and parameters: {'bagging_fraction': 0.9933366847840979, 'bagging_freq': 7}. Best is trial 27 with value: 0.6900113794489978.
bagging, val_score: 0.689984: 40%|#### | 4/10 [00:03<00:05, 1.02it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001593 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663295 valid's binary_logloss: 0.692575
Early stopping, best iteration is:
[46] train's binary_logloss: 0.676198 valid's binary_logloss: 0.690885
bagging, val_score: 0.689984: 50%|##### | 5/10 [00:04<00:04, 1.04it/s][I 2020-09-27 04:57:39,840] Trial 31 finished with value: 0.6908846966555133 and parameters: {'bagging_fraction': 0.7357347346451741, 'bagging_freq': 4}. Best is trial 27 with value: 0.6900113794489978.
bagging, val_score: 0.689984: 50%|##### | 5/10 [00:04<00:04, 1.04it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001590 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.662926 valid's binary_logloss: 0.691444
Early stopping, best iteration is:
[22] train's binary_logloss: 0.683062 valid's binary_logloss: 0.690387
bagging, val_score: 0.689984: 60%|###### | 6/10 [00:05<00:03, 1.09it/s][I 2020-09-27 04:57:40,659] Trial 32 finished with value: 0.6903865124235224 and parameters: {'bagging_fraction': 0.9606439169637138, 'bagging_freq': 4}. Best is trial 27 with value: 0.6900113794489978.
bagging, val_score: 0.689984: 60%|###### | 6/10 [00:05<00:03, 1.09it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001535 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.66338 valid's binary_logloss: 0.692174
Early stopping, best iteration is:
[35] train's binary_logloss: 0.679143 valid's binary_logloss: 0.690222
bagging, val_score: 0.689984: 70%|####### | 7/10 [00:06<00:02, 1.15it/s][I 2020-09-27 04:57:41,428] Trial 33 finished with value: 0.6902221405606176 and parameters: {'bagging_fraction': 0.5501404306458005, 'bagging_freq': 1}. Best is trial 27 with value: 0.6900113794489978.
bagging, val_score: 0.689984: 70%|####### | 7/10 [00:06<00:02, 1.15it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001493 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.662525 valid's binary_logloss: 0.691554
Early stopping, best iteration is:
[30] train's binary_logloss: 0.680411 valid's binary_logloss: 0.690703
bagging, val_score: 0.689984: 80%|######## | 8/10 [00:07<00:01, 1.16it/s][I 2020-09-27 04:57:42,266] Trial 34 finished with value: 0.6907033298544214 and parameters: {'bagging_fraction': 0.8542206886051794, 'bagging_freq': 6}. Best is trial 27 with value: 0.6900113794489978.
bagging, val_score: 0.689984: 80%|######## | 8/10 [00:07<00:01, 1.16it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011907 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663364 valid's binary_logloss: 0.69162
Early stopping, best iteration is:
[35] train's binary_logloss: 0.679523 valid's binary_logloss: 0.690026
bagging, val_score: 0.689984: 90%|######### | 9/10 [00:08<00:00, 1.18it/s][I 2020-09-27 04:57:43,075] Trial 35 finished with value: 0.6900257534785464 and parameters: {'bagging_fraction': 0.6072769654268587, 'bagging_freq': 3}. Best is trial 27 with value: 0.6900113794489978.
bagging, val_score: 0.689984: 90%|######### | 9/10 [00:08<00:00, 1.18it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001565 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663332 valid's binary_logloss: 0.691011
Early stopping, best iteration is:
[30] train's binary_logloss: 0.680985 valid's binary_logloss: 0.69027
bagging, val_score: 0.689984: 100%|##########| 10/10 [00:08<00:00, 1.19it/s][I 2020-09-27 04:57:43,910] Trial 36 finished with value: 0.6902695334181121 and parameters: {'bagging_fraction': 0.618007946081386, 'bagging_freq': 3}. Best is trial 27 with value: 0.6900113794489978.
bagging, val_score: 0.689984: 100%|##########| 10/10 [00:08<00:00, 1.11it/s]
feature_fraction_stage2, val_score: 0.689984: 0%| | 0/6 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013405 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.662852 valid's binary_logloss: 0.691261
Early stopping, best iteration is:
[51] train's binary_logloss: 0.674585 valid's binary_logloss: 0.690386
feature_fraction_stage2, val_score: 0.689984: 17%|#6 | 1/6 [00:00<00:04, 1.23it/s][I 2020-09-27 04:57:44,733] Trial 37 finished with value: 0.690385946557375 and parameters: {'feature_fraction': 0.7520000000000001}. Best is trial 37 with value: 0.690385946557375.
feature_fraction_stage2, val_score: 0.689984: 17%|#6 | 1/6 [00:00<00:04, 1.23it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008459 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663021 valid's binary_logloss: 0.690361
Early stopping, best iteration is:
[48] train's binary_logloss: 0.675428 valid's binary_logloss: 0.689984
feature_fraction_stage2, val_score: 0.689984: 33%|###3 | 2/6 [00:01<00:03, 1.23it/s][I 2020-09-27 04:57:45,549] Trial 38 finished with value: 0.6899844980770151 and parameters: {'feature_fraction': 0.8160000000000001}. Best is trial 38 with value: 0.6899844980770151.
feature_fraction_stage2, val_score: 0.689984: 33%|###3 | 2/6 [00:01<00:03, 1.23it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001668 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.662759 valid's binary_logloss: 0.691799
Early stopping, best iteration is:
[36] train's binary_logloss: 0.678598 valid's binary_logloss: 0.69056
feature_fraction_stage2, val_score: 0.689984: 50%|##### | 3/6 [00:02<00:02, 1.26it/s][I 2020-09-27 04:57:46,293] Trial 39 finished with value: 0.6905601497910684 and parameters: {'feature_fraction': 0.88}. Best is trial 38 with value: 0.6899844980770151.
feature_fraction_stage2, val_score: 0.689984: 50%|##### | 3/6 [00:02<00:02, 1.26it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008728 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.662852 valid's binary_logloss: 0.691261
Early stopping, best iteration is:
[51] train's binary_logloss: 0.674585 valid's binary_logloss: 0.690386
feature_fraction_stage2, val_score: 0.689984: 67%|######6 | 4/6 [00:03<00:01, 1.27it/s][I 2020-09-27 04:57:47,065] Trial 40 finished with value: 0.6903859465573751 and parameters: {'feature_fraction': 0.784}. Best is trial 38 with value: 0.6899844980770151.
feature_fraction_stage2, val_score: 0.689984: 67%|######6 | 4/6 [00:03<00:01, 1.27it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013482 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663303 valid's binary_logloss: 0.690672
Early stopping, best iteration is:
[53] train's binary_logloss: 0.674298 valid's binary_logloss: 0.690024
feature_fraction_stage2, val_score: 0.689984: 83%|########3 | 5/6 [00:03<00:00, 1.27it/s][I 2020-09-27 04:57:47,852] Trial 41 finished with value: 0.6900236664988709 and parameters: {'feature_fraction': 0.7200000000000001}. Best is trial 38 with value: 0.6899844980770151.
feature_fraction_stage2, val_score: 0.689984: 83%|########3 | 5/6 [00:03<00:00, 1.27it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009074 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.662643 valid's binary_logloss: 0.690983
Early stopping, best iteration is:
[32] train's binary_logloss: 0.679699 valid's binary_logloss: 0.690557
feature_fraction_stage2, val_score: 0.689984: 100%|##########| 6/6 [00:04<00:00, 1.31it/s][I 2020-09-27 04:57:48,572] Trial 42 finished with value: 0.6905568387008639 and parameters: {'feature_fraction': 0.8480000000000001}. Best is trial 38 with value: 0.6899844980770151.
feature_fraction_stage2, val_score: 0.689984: 100%|##########| 6/6 [00:04<00:00, 1.29it/s]
regularization_factors, val_score: 0.689984: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001489 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663071 valid's binary_logloss: 0.690424
Early stopping, best iteration is:
[79] train's binary_logloss: 0.667888 valid's binary_logloss: 0.689968
regularization_factors, val_score: 0.689968: 5%|5 | 1/20 [00:01<00:19, 1.00s/it][I 2020-09-27 04:57:49,589] Trial 43 finished with value: 0.6899675224070905 and parameters: {'lambda_l1': 0.0034339426266632865, 'lambda_l2': 2.9528148173408035e-06}. Best is trial 43 with value: 0.6899675224070905.
regularization_factors, val_score: 0.689968: 5%|5 | 1/20 [00:01<00:19, 1.00s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001532 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.66306 valid's binary_logloss: 0.690447
Early stopping, best iteration is:
[79] train's binary_logloss: 0.66789 valid's binary_logloss: 0.689967
regularization_factors, val_score: 0.689967: 10%|# | 2/20 [00:02<00:18, 1.01s/it][I 2020-09-27 04:57:50,613] Trial 44 finished with value: 0.6899674411707333 and parameters: {'lambda_l1': 0.004662561477422725, 'lambda_l2': 1.3948625139125175e-06}. Best is trial 44 with value: 0.6899674411707333.
regularization_factors, val_score: 0.689967: 10%|# | 2/20 [00:02<00:18, 1.01s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.003785 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663061 valid's binary_logloss: 0.690447
Early stopping, best iteration is:
[79] train's binary_logloss: 0.66789 valid's binary_logloss: 0.689967
regularization_factors, val_score: 0.689967: 15%|#5 | 3/20 [00:03<00:17, 1.01s/it][I 2020-09-27 04:57:51,643] Trial 45 finished with value: 0.6899674025098815 and parameters: {'lambda_l1': 0.00524640370808824, 'lambda_l2': 1.4398621808620703e-06}. Best is trial 45 with value: 0.6899674025098815.
regularization_factors, val_score: 0.689967: 15%|#5 | 3/20 [00:03<00:17, 1.01s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008016 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.662786 valid's binary_logloss: 0.690799
Early stopping, best iteration is:
[47] train's binary_logloss: 0.67568 valid's binary_logloss: 0.69003
regularization_factors, val_score: 0.689967: 20%|## | 4/20 [00:03<00:15, 1.05it/s][I 2020-09-27 04:57:52,442] Trial 46 finished with value: 0.6900296520521778 and parameters: {'lambda_l1': 0.010031342307687478, 'lambda_l2': 8.690635159166719e-07}. Best is trial 45 with value: 0.6899674025098815.
regularization_factors, val_score: 0.689967: 20%|## | 4/20 [00:03<00:15, 1.05it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001557 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.66307 valid's binary_logloss: 0.690424
Early stopping, best iteration is:
[79] train's binary_logloss: 0.667887 valid's binary_logloss: 0.689968
regularization_factors, val_score: 0.689967: 25%|##5 | 5/20 [00:04<00:14, 1.05it/s][I 2020-09-27 04:57:53,400] Trial 47 finished with value: 0.6899675786092242 and parameters: {'lambda_l1': 0.002586041944855055, 'lambda_l2': 2.63283121364215e-06}. Best is trial 45 with value: 0.6899674025098815.
regularization_factors, val_score: 0.689967: 25%|##5 | 5/20 [00:04<00:14, 1.05it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008047 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663059 valid's binary_logloss: 0.690447
Early stopping, best iteration is:
[79] train's binary_logloss: 0.667889 valid's binary_logloss: 0.689967
regularization_factors, val_score: 0.689967: 30%|### | 6/20 [00:05<00:13, 1.05it/s][I 2020-09-27 04:57:54,339] Trial 48 finished with value: 0.6899674839198124 and parameters: {'lambda_l1': 0.004015863099526828, 'lambda_l2': 2.280356884231991e-06}. Best is trial 45 with value: 0.6899674025098815.
regularization_factors, val_score: 0.689967: 30%|### | 6/20 [00:05<00:13, 1.05it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001644 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663071 valid's binary_logloss: 0.690424
Early stopping, best iteration is:
[79] train's binary_logloss: 0.667888 valid's binary_logloss: 0.689968
regularization_factors, val_score: 0.689967: 35%|###5 | 7/20 [00:06<00:12, 1.01it/s][I 2020-09-27 04:57:55,418] Trial 49 finished with value: 0.6899675288047956 and parameters: {'lambda_l1': 0.0033375593376080356, 'lambda_l2': 2.830642312376846e-06}. Best is trial 45 with value: 0.6899674025098815.
regularization_factors, val_score: 0.689967: 35%|###5 | 7/20 [00:06<00:12, 1.01it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001601 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663071 valid's binary_logloss: 0.690424
Early stopping, best iteration is:
[79] train's binary_logloss: 0.667888 valid's binary_logloss: 0.689968
regularization_factors, val_score: 0.689967: 40%|#### | 8/20 [00:07<00:12, 1.00s/it][I 2020-09-27 04:57:56,448] Trial 50 finished with value: 0.6899675296729249 and parameters: {'lambda_l1': 0.0033245647295823137, 'lambda_l2': 2.7333945263234923e-06}. Best is trial 45 with value: 0.6899674025098815.
regularization_factors, val_score: 0.689967: 40%|#### | 8/20 [00:07<00:12, 1.00s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001603 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663071 valid's binary_logloss: 0.690424
Early stopping, best iteration is:
[79] train's binary_logloss: 0.667888 valid's binary_logloss: 0.689968
regularization_factors, val_score: 0.689967: 45%|####5 | 9/20 [00:08<00:10, 1.01it/s][I 2020-09-27 04:57:57,415] Trial 51 finished with value: 0.6899675231589405 and parameters: {'lambda_l1': 0.0034231822486881607, 'lambda_l2': 2.5597586648530185e-06}. Best is trial 45 with value: 0.6899674025098815.
regularization_factors, val_score: 0.689967: 45%|####5 | 9/20 [00:08<00:10, 1.01it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001568 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663071 valid's binary_logloss: 0.690424
Early stopping, best iteration is:
[79] train's binary_logloss: 0.667888 valid's binary_logloss: 0.689968
regularization_factors, val_score: 0.689967: 50%|##### | 10/20 [00:09<00:09, 1.02it/s][I 2020-09-27 04:57:58,387] Trial 52 finished with value: 0.6899675209429271 and parameters: {'lambda_l1': 0.0034569782502191274, 'lambda_l2': 2.163106922428234e-06}. Best is trial 45 with value: 0.6899674025098815.
regularization_factors, val_score: 0.689967: 50%|##### | 10/20 [00:09<00:09, 1.02it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001571 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663071 valid's binary_logloss: 0.690424
Early stopping, best iteration is:
[79] train's binary_logloss: 0.667889 valid's binary_logloss: 0.689968
regularization_factors, val_score: 0.689967: 55%|#####5 | 11/20 [00:10<00:08, 1.02it/s][I 2020-09-27 04:57:59,369] Trial 53 finished with value: 0.6899675043589648 and parameters: {'lambda_l1': 0.003707901920209718, 'lambda_l2': 1.70752737693983e-06}. Best is trial 45 with value: 0.6899674025098815.
regularization_factors, val_score: 0.689967: 55%|#####5 | 11/20 [00:10<00:08, 1.02it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001536 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.66306 valid's binary_logloss: 0.690447
Early stopping, best iteration is:
[79] train's binary_logloss: 0.667889 valid's binary_logloss: 0.689967
regularization_factors, val_score: 0.689967: 60%|###### | 12/20 [00:11<00:07, 1.01it/s][I 2020-09-27 04:58:00,373] Trial 54 finished with value: 0.6899674606296653 and parameters: {'lambda_l1': 0.004368898887817153, 'lambda_l2': 1.2458711221699426e-06}. Best is trial 45 with value: 0.6899674025098815.
regularization_factors, val_score: 0.689967: 60%|###### | 12/20 [00:11<00:07, 1.01it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001596 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.662937 valid's binary_logloss: 0.690557
Early stopping, best iteration is:
[47] train's binary_logloss: 0.675852 valid's binary_logloss: 0.690271
regularization_factors, val_score: 0.689967: 65%|######5 | 13/20 [00:12<00:06, 1.05it/s][I 2020-09-27 04:58:01,228] Trial 55 finished with value: 0.6902710837967893 and parameters: {'lambda_l1': 0.07746158051498105, 'lambda_l2': 3.231480194675918e-08}. Best is trial 45 with value: 0.6899674025098815.
regularization_factors, val_score: 0.689967: 65%|######5 | 13/20 [00:12<00:06, 1.05it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008971 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664683 valid's binary_logloss: 0.690143
Early stopping, best iteration is:
[52] train's binary_logloss: 0.675081 valid's binary_logloss: 0.689647
regularization_factors, val_score: 0.689647: 70%|####### | 14/20 [00:13<00:05, 1.08it/s][I 2020-09-27 04:58:02,088] Trial 56 finished with value: 0.6896473829748281 and parameters: {'lambda_l1': 5.946750740071561e-06, 'lambda_l2': 2.6997857156008718}. Best is trial 56 with value: 0.6896473829748281.
regularization_factors, val_score: 0.689647: 70%|####### | 14/20 [00:13<00:05, 1.08it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001490 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665021 valid's binary_logloss: 0.691158
Early stopping, best iteration is:
[47] train's binary_logloss: 0.676401 valid's binary_logloss: 0.6902
regularization_factors, val_score: 0.689647: 75%|#######5 | 15/20 [00:14<00:04, 1.11it/s][I 2020-09-27 04:58:02,943] Trial 57 finished with value: 0.6902003315084454 and parameters: {'lambda_l1': 5.758495952716345e-07, 'lambda_l2': 3.755109192731262}. Best is trial 56 with value: 0.6896473829748281.
regularization_factors, val_score: 0.689647: 75%|#######5 | 15/20 [00:14<00:04, 1.11it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001569 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663022 valid's binary_logloss: 0.690361
Early stopping, best iteration is:
[48] train's binary_logloss: 0.675429 valid's binary_logloss: 0.689984
regularization_factors, val_score: 0.689647: 80%|######## | 16/20 [00:15<00:03, 1.14it/s][I 2020-09-27 04:58:03,771] Trial 58 finished with value: 0.6899844790092887 and parameters: {'lambda_l1': 2.0582999452425637e-05, 'lambda_l2': 0.0005026756920447598}. Best is trial 56 with value: 0.6896473829748281.
regularization_factors, val_score: 0.689647: 80%|######## | 16/20 [00:15<00:03, 1.14it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001565 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663021 valid's binary_logloss: 0.690361
Early stopping, best iteration is:
[48] train's binary_logloss: 0.675428 valid's binary_logloss: 0.689984
regularization_factors, val_score: 0.689647: 85%|########5 | 17/20 [00:16<00:02, 1.16it/s][I 2020-09-27 04:58:04,597] Trial 59 finished with value: 0.6899844953569758 and parameters: {'lambda_l1': 6.0992831053207785e-05, 'lambda_l2': 4.252592969567293e-08}. Best is trial 56 with value: 0.6896473829748281.
regularization_factors, val_score: 0.689647: 85%|########5 | 17/20 [00:16<00:02, 1.16it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001929 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663089 valid's binary_logloss: 0.691109
Early stopping, best iteration is:
[43] train's binary_logloss: 0.676731 valid's binary_logloss: 0.690043
regularization_factors, val_score: 0.689647: 90%|######### | 18/20 [00:16<00:01, 1.17it/s][I 2020-09-27 04:58:05,432] Trial 60 finished with value: 0.6900427743024334 and parameters: {'lambda_l1': 0.268567393042738, 'lambda_l2': 0.0003824677053871511}. Best is trial 56 with value: 0.6896473829748281.
regularization_factors, val_score: 0.689647: 90%|######### | 18/20 [00:16<00:01, 1.17it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001538 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663047 valid's binary_logloss: 0.690658
Early stopping, best iteration is:
[48] train's binary_logloss: 0.675429 valid's binary_logloss: 0.689984
regularization_factors, val_score: 0.689647: 95%|#########5| 19/20 [00:17<00:00, 1.17it/s][I 2020-09-27 04:58:06,282] Trial 61 finished with value: 0.6899844746373079 and parameters: {'lambda_l1': 0.0005260392810081732, 'lambda_l2': 2.7967911031858336e-07}. Best is trial 56 with value: 0.6896473829748281.
regularization_factors, val_score: 0.689647: 95%|#########5| 19/20 [00:17<00:00, 1.17it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001556 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663199 valid's binary_logloss: 0.690198
Early stopping, best iteration is:
[52] train's binary_logloss: 0.674513 valid's binary_logloss: 0.68975
regularization_factors, val_score: 0.689647: 100%|##########| 20/20 [00:18<00:00, 1.15it/s][I 2020-09-27 04:58:07,180] Trial 62 finished with value: 0.6897502601908707 and parameters: {'lambda_l1': 0.03166566393209921, 'lambda_l2': 3.6505882507400447e-05}. Best is trial 56 with value: 0.6896473829748281.
regularization_factors, val_score: 0.689647: 100%|##########| 20/20 [00:18<00:00, 1.08it/s]
min_data_in_leaf, val_score: 0.689647: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011144 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664472 valid's binary_logloss: 0.690172
[200] train's binary_logloss: 0.645537 valid's binary_logloss: 0.690997
Early stopping, best iteration is:
[108] train's binary_logloss: 0.662894 valid's binary_logloss: 0.690057
min_data_in_leaf, val_score: 0.689647: 20%|## | 1/5 [00:01<00:04, 1.19s/it][I 2020-09-27 04:58:08,386] Trial 63 finished with value: 0.6900565147753162 and parameters: {'min_child_samples': 10}. Best is trial 63 with value: 0.6900565147753162.
min_data_in_leaf, val_score: 0.689647: 20%|## | 1/5 [00:01<00:04, 1.19s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002284 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665148 valid's binary_logloss: 0.691028
Early stopping, best iteration is:
[49] train's binary_logloss: 0.676081 valid's binary_logloss: 0.690354
min_data_in_leaf, val_score: 0.689647: 40%|#### | 2/5 [00:02<00:03, 1.10s/it][I 2020-09-27 04:58:09,276] Trial 64 finished with value: 0.6903537539012076 and parameters: {'min_child_samples': 100}. Best is trial 63 with value: 0.6900565147753162.
min_data_in_leaf, val_score: 0.689647: 40%|#### | 2/5 [00:02<00:03, 1.10s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007936 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664812 valid's binary_logloss: 0.690985
Early stopping, best iteration is:
[41] train's binary_logloss: 0.677793 valid's binary_logloss: 0.690549
min_data_in_leaf, val_score: 0.689647: 60%|###### | 3/5 [00:02<00:02, 1.02s/it][I 2020-09-27 04:58:10,108] Trial 65 finished with value: 0.690549034445801 and parameters: {'min_child_samples': 5}. Best is trial 63 with value: 0.6900565147753162.
min_data_in_leaf, val_score: 0.689647: 60%|###### | 3/5 [00:02<00:02, 1.02s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001560 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664729 valid's binary_logloss: 0.690892
Early stopping, best iteration is:
[41] train's binary_logloss: 0.677757 valid's binary_logloss: 0.690151
min_data_in_leaf, val_score: 0.689647: 80%|######## | 4/5 [00:03<00:00, 1.03it/s][I 2020-09-27 04:58:10,959] Trial 66 finished with value: 0.6901508243301284 and parameters: {'min_child_samples': 25}. Best is trial 63 with value: 0.6900565147753162.
min_data_in_leaf, val_score: 0.689647: 80%|######## | 4/5 [00:03<00:00, 1.03it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001581 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128
[LightGBM] [Info] Start training from score -0.004128
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664643 valid's binary_logloss: 0.691412
Early stopping, best iteration is:
[39] train's binary_logloss: 0.678314 valid's binary_logloss: 0.690208
min_data_in_leaf, val_score: 0.689647: 100%|##########| 5/5 [00:04<00:00, 1.07it/s][I 2020-09-27 04:58:11,803] Trial 67 finished with value: 0.6902082945895022 and parameters: {'min_child_samples': 50}. Best is trial 63 with value: 0.6900565147753162.
min_data_in_leaf, val_score: 0.689647: 100%|##########| 5/5 [00:04<00:00, 1.08it/s]
Fold : 6
[I 2020-09-27 04:58:11,870] A new study created in memory with name: no-name-dcbdec6e-2ceb-4d1e-86b5-bfbd5414d073
feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000995 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665269 valid's binary_logloss: 0.688894
[200] train's binary_logloss: 0.645599 valid's binary_logloss: 0.690352
Early stopping, best iteration is:
[106] train's binary_logloss: 0.663937 valid's binary_logloss: 0.688771
feature_fraction, val_score: 0.688771: 14%|#4 | 1/7 [00:01<00:07, 1.19s/it][I 2020-09-27 04:58:13,071] Trial 0 finished with value: 0.6887706335922763 and parameters: {'feature_fraction': 0.5}. Best is trial 0 with value: 0.6887706335922763.
feature_fraction, val_score: 0.688771: 14%|#4 | 1/7 [00:01<00:07, 1.19s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001614 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.662986 valid's binary_logloss: 0.689796
Early stopping, best iteration is:
[32] train's binary_logloss: 0.679798 valid's binary_logloss: 0.689557
feature_fraction, val_score: 0.688771: 29%|##8 | 2/7 [00:01<00:05, 1.06s/it][I 2020-09-27 04:58:13,811] Trial 1 finished with value: 0.689557320235649 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 0 with value: 0.6887706335922763.
feature_fraction, val_score: 0.688771: 29%|##8 | 2/7 [00:01<00:05, 1.06s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.015363 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663698 valid's binary_logloss: 0.689777
Early stopping, best iteration is:
[62] train's binary_logloss: 0.672409 valid's binary_logloss: 0.689451
feature_fraction, val_score: 0.688771: 43%|####2 | 3/7 [00:02<00:03, 1.02it/s][I 2020-09-27 04:58:14,610] Trial 2 finished with value: 0.6894514131827376 and parameters: {'feature_fraction': 0.7}. Best is trial 0 with value: 0.6887706335922763.
feature_fraction, val_score: 0.688771: 43%|####2 | 3/7 [00:02<00:03, 1.02it/s][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007904 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.66431 valid's binary_logloss: 0.689418
Early stopping, best iteration is:
[82] train's binary_logloss: 0.668145 valid's binary_logloss: 0.689215
feature_fraction, val_score: 0.688771: 57%|#####7 | 4/7 [00:03<00:02, 1.07it/s][I 2020-09-27 04:58:15,452] Trial 3 finished with value: 0.6892151532007572 and parameters: {'feature_fraction': 0.6}. Best is trial 0 with value: 0.6887706335922763.
feature_fraction, val_score: 0.688771: 57%|#####7 | 4/7 [00:03<00:02, 1.07it/s][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008838 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663214 valid's binary_logloss: 0.689646
Early stopping, best iteration is:
[61] train's binary_logloss: 0.672363 valid's binary_logloss: 0.689118
feature_fraction, val_score: 0.688771: 71%|#######1 | 5/7 [00:04<00:01, 1.11it/s][I 2020-09-27 04:58:16,267] Trial 4 finished with value: 0.6891176758344856 and parameters: {'feature_fraction': 0.8}. Best is trial 0 with value: 0.6887706335922763.
feature_fraction, val_score: 0.688771: 71%|#######1 | 5/7 [00:04<00:01, 1.11it/s][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000778 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.66613 valid's binary_logloss: 0.689486
Early stopping, best iteration is:
[75] train's binary_logloss: 0.671175 valid's binary_logloss: 0.689183
feature_fraction, val_score: 0.688771: 86%|########5 | 6/7 [00:05<00:00, 1.13it/s][I 2020-09-27 04:58:17,113] Trial 5 finished with value: 0.6891830705136953 and parameters: {'feature_fraction': 0.4}. Best is trial 0 with value: 0.6887706335922763.
feature_fraction, val_score: 0.688771: 86%|########5 | 6/7 [00:05<00:00, 1.13it/s][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001832 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.662507 valid's binary_logloss: 0.690168
Early stopping, best iteration is:
[48] train's binary_logloss: 0.675204 valid's binary_logloss: 0.689566
feature_fraction, val_score: 0.688771: 100%|##########| 7/7 [00:07<00:00, 1.32s/it][I 2020-09-27 04:58:19,455] Trial 6 finished with value: 0.6895663490864646 and parameters: {'feature_fraction': 1.0}. Best is trial 0 with value: 0.6887706335922763.
feature_fraction, val_score: 0.688771: 100%|##########| 7/7 [00:07<00:00, 1.08s/it]
num_leaves, val_score: 0.688771: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000875 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.65635 valid's binary_logloss: 0.690173
Early stopping, best iteration is:
[43] train's binary_logloss: 0.673607 valid's binary_logloss: 0.689502
num_leaves, val_score: 0.688771: 5%|5 | 1/20 [00:01<00:20, 1.08s/it][I 2020-09-27 04:58:20,546] Trial 7 finished with value: 0.6895015997019708 and parameters: {'num_leaves': 43}. Best is trial 7 with value: 0.6895015997019708.
num_leaves, val_score: 0.688771: 5%|5 | 1/20 [00:01<00:20, 1.08s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000948 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.636388 valid's binary_logloss: 0.690447
Early stopping, best iteration is:
[88] train's binary_logloss: 0.641912 valid's binary_logloss: 0.689924
num_leaves, val_score: 0.688771: 10%|# | 2/20 [00:02<00:20, 1.13s/it][I 2020-09-27 04:58:21,813] Trial 8 finished with value: 0.6899239348271381 and parameters: {'num_leaves': 73}. Best is trial 7 with value: 0.6895015997019708.
num_leaves, val_score: 0.688771: 10%|# | 2/20 [00:02<00:20, 1.13s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001001 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.640263 valid's binary_logloss: 0.690357
Early stopping, best iteration is:
[50] train's binary_logloss: 0.661952 valid's binary_logloss: 0.689585
num_leaves, val_score: 0.688771: 15%|#5 | 3/20 [00:03<00:18, 1.10s/it][I 2020-09-27 04:58:22,822] Trial 9 finished with value: 0.6895847174318764 and parameters: {'num_leaves': 67}. Best is trial 7 with value: 0.6895015997019708.
num_leaves, val_score: 0.688771: 15%|#5 | 3/20 [00:03<00:18, 1.10s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001688 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.552597 valid's binary_logloss: 0.69652
Early stopping, best iteration is:
[17] train's binary_logloss: 0.65881 valid's binary_logloss: 0.692576
num_leaves, val_score: 0.688771: 20%|## | 4/20 [00:04<00:19, 1.19s/it][I 2020-09-27 04:58:24,227] Trial 10 finished with value: 0.6925759831330955 and parameters: {'num_leaves': 228}. Best is trial 7 with value: 0.6895015997019708.
num_leaves, val_score: 0.688771: 20%|## | 4/20 [00:04<00:19, 1.19s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001533 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.538026 valid's binary_logloss: 0.69752
Early stopping, best iteration is:
[15] train's binary_logloss: 0.659639 valid's binary_logloss: 0.691473
num_leaves, val_score: 0.688771: 25%|##5 | 5/20 [00:06<00:18, 1.25s/it][I 2020-09-27 04:58:25,631] Trial 11 finished with value: 0.6914732359891139 and parameters: {'num_leaves': 254}. Best is trial 7 with value: 0.6895015997019708.
num_leaves, val_score: 0.688771: 25%|##5 | 5/20 [00:06<00:18, 1.25s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001108 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.578766 valid's binary_logloss: 0.694226
Early stopping, best iteration is:
[25] train's binary_logloss: 0.654454 valid's binary_logloss: 0.691199
num_leaves, val_score: 0.688771: 30%|### | 6/20 [00:07<00:17, 1.22s/it][I 2020-09-27 04:58:26,778] Trial 12 finished with value: 0.6911992488401045 and parameters: {'num_leaves': 174}. Best is trial 7 with value: 0.6895015997019708.
num_leaves, val_score: 0.688771: 30%|### | 6/20 [00:07<00:17, 1.22s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000950 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.587943 valid's binary_logloss: 0.69436
Early stopping, best iteration is:
[64] train's binary_logloss: 0.61787 valid's binary_logloss: 0.691342
num_leaves, val_score: 0.688771: 35%|###5 | 7/20 [00:08<00:16, 1.27s/it][I 2020-09-27 04:58:28,162] Trial 13 finished with value: 0.6913417018956572 and parameters: {'num_leaves': 156}. Best is trial 7 with value: 0.6895015997019708.
num_leaves, val_score: 0.688771: 35%|###5 | 7/20 [00:08<00:16, 1.27s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001185 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.675867 valid's binary_logloss: 0.689217
[200] train's binary_logloss: 0.664751 valid's binary_logloss: 0.689817
Early stopping, best iteration is:
[111] train's binary_logloss: 0.674533 valid's binary_logloss: 0.68909
num_leaves, val_score: 0.688771: 40%|#### | 8/20 [00:09<00:14, 1.18s/it][I 2020-09-27 04:58:29,127] Trial 14 finished with value: 0.6890903374716104 and parameters: {'num_leaves': 17}. Best is trial 14 with value: 0.6890903374716104.
num_leaves, val_score: 0.688771: 40%|#### | 8/20 [00:09<00:14, 1.18s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001614 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688527 valid's binary_logloss: 0.690133
[200] train's binary_logloss: 0.686734 valid's binary_logloss: 0.689033
[300] train's binary_logloss: 0.685468 valid's binary_logloss: 0.688641
[400] train's binary_logloss: 0.684369 valid's binary_logloss: 0.688513
Early stopping, best iteration is:
[375] train's binary_logloss: 0.684633 valid's binary_logloss: 0.68848
num_leaves, val_score: 0.688480: 45%|####5 | 9/20 [00:11<00:14, 1.32s/it][I 2020-09-27 04:58:30,769] Trial 15 finished with value: 0.6884795876194431 and parameters: {'num_leaves': 3}. Best is trial 15 with value: 0.6884795876194431.
num_leaves, val_score: 0.688480: 45%|####5 | 9/20 [00:11<00:14, 1.32s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001058 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.610458 valid's binary_logloss: 0.69108
Early stopping, best iteration is:
[49] train's binary_logloss: 0.645426 valid's binary_logloss: 0.690053
num_leaves, val_score: 0.688480: 50%|##### | 10/20 [00:12<00:12, 1.27s/it][I 2020-09-27 04:58:31,926] Trial 16 finished with value: 0.6900527437374864 and parameters: {'num_leaves': 115}. Best is trial 15 with value: 0.6884795876194431.
num_leaves, val_score: 0.688480: 50%|##### | 10/20 [00:12<00:12, 1.27s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000997 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688527 valid's binary_logloss: 0.690133
[200] train's binary_logloss: 0.686734 valid's binary_logloss: 0.689033
[300] train's binary_logloss: 0.685468 valid's binary_logloss: 0.688641
[400] train's binary_logloss: 0.684369 valid's binary_logloss: 0.688513
Early stopping, best iteration is:
[375] train's binary_logloss: 0.684633 valid's binary_logloss: 0.68848
num_leaves, val_score: 0.688480: 55%|#####5 | 11/20 [00:14<00:12, 1.37s/it][I 2020-09-27 04:58:33,539] Trial 17 finished with value: 0.6884795876194431 and parameters: {'num_leaves': 3}. Best is trial 15 with value: 0.6884795876194431.
num_leaves, val_score: 0.688480: 55%|#####5 | 11/20 [00:14<00:12, 1.37s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000988 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.690025 valid's binary_logloss: 0.691003
[200] train's binary_logloss: 0.688901 valid's binary_logloss: 0.69002
[300] train's binary_logloss: 0.688185 valid's binary_logloss: 0.68941
[400] train's binary_logloss: 0.687683 valid's binary_logloss: 0.688943
[500] train's binary_logloss: 0.687304 valid's binary_logloss: 0.688654
[600] train's binary_logloss: 0.687004 valid's binary_logloss: 0.688502
[700] train's binary_logloss: 0.686761 valid's binary_logloss: 0.688428
[800] train's binary_logloss: 0.686556 valid's binary_logloss: 0.688335
[900] train's binary_logloss: 0.68638 valid's binary_logloss: 0.688325
[1000] train's binary_logloss: 0.686227 valid's binary_logloss: 0.688255
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.686227 valid's binary_logloss: 0.688255
num_leaves, val_score: 0.688255: 60%|###### | 12/20 [00:17<00:15, 1.92s/it][I 2020-09-27 04:58:36,728] Trial 18 finished with value: 0.6882551844633495 and parameters: {'num_leaves': 2}. Best is trial 18 with value: 0.6882551844633495.
num_leaves, val_score: 0.688255: 60%|###### | 12/20 [00:17<00:15, 1.92s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001002 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.683515 valid's binary_logloss: 0.689389
[200] train's binary_logloss: 0.678306 valid's binary_logloss: 0.689315
[300] train's binary_logloss: 0.673658 valid's binary_logloss: 0.688984
[400] train's binary_logloss: 0.669291 valid's binary_logloss: 0.689314
Early stopping, best iteration is:
[317] train's binary_logloss: 0.6729 valid's binary_logloss: 0.688929
num_leaves, val_score: 0.688255: 65%|######5 | 13/20 [00:18<00:12, 1.81s/it][I 2020-09-27 04:58:38,296] Trial 19 finished with value: 0.68892857284493 and parameters: {'num_leaves': 8}. Best is trial 18 with value: 0.6882551844633495.
num_leaves, val_score: 0.688255: 65%|######5 | 13/20 [00:18<00:12, 1.81s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000915 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.62055 valid's binary_logloss: 0.692044
Early stopping, best iteration is:
[29] train's binary_logloss: 0.66542 valid's binary_logloss: 0.690281
num_leaves, val_score: 0.688255: 70%|####### | 14/20 [00:19<00:09, 1.55s/it][I 2020-09-27 04:58:39,243] Trial 20 finished with value: 0.6902812591323769 and parameters: {'num_leaves': 99}. Best is trial 18 with value: 0.6882551844633495.
num_leaves, val_score: 0.688255: 70%|####### | 14/20 [00:19<00:09, 1.55s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001482 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.684441 valid's binary_logloss: 0.689008
[200] train's binary_logloss: 0.679968 valid's binary_logloss: 0.688832
[300] train's binary_logloss: 0.676016 valid's binary_logloss: 0.688309
[400] train's binary_logloss: 0.672203 valid's binary_logloss: 0.688492
Early stopping, best iteration is:
[301] train's binary_logloss: 0.675968 valid's binary_logloss: 0.688303
num_leaves, val_score: 0.688255: 75%|#######5 | 15/20 [00:21<00:07, 1.52s/it][I 2020-09-27 04:58:40,689] Trial 21 finished with value: 0.6883028732185514 and parameters: {'num_leaves': 7}. Best is trial 18 with value: 0.6882551844633495.
num_leaves, val_score: 0.688255: 75%|#######5 | 15/20 [00:21<00:07, 1.52s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000968 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.662251 valid's binary_logloss: 0.689855
Early stopping, best iteration is:
[71] train's binary_logloss: 0.669171 valid's binary_logloss: 0.689288
num_leaves, val_score: 0.688255: 80%|######## | 16/20 [00:22<00:05, 1.33s/it][I 2020-09-27 04:58:41,585] Trial 22 finished with value: 0.6892878269668582 and parameters: {'num_leaves': 35}. Best is trial 18 with value: 0.6882551844633495.
num_leaves, val_score: 0.688255: 80%|######## | 16/20 [00:22<00:05, 1.33s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000967 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68536 valid's binary_logloss: 0.689552
[200] train's binary_logloss: 0.681488 valid's binary_logloss: 0.689209
[300] train's binary_logloss: 0.678139 valid's binary_logloss: 0.68914
Early stopping, best iteration is:
[279] train's binary_logloss: 0.67879 valid's binary_logloss: 0.689
num_leaves, val_score: 0.688255: 85%|########5 | 17/20 [00:23<00:04, 1.35s/it][I 2020-09-27 04:58:42,989] Trial 23 finished with value: 0.6889999863028395 and parameters: {'num_leaves': 6}. Best is trial 18 with value: 0.6882551844633495.
num_leaves, val_score: 0.688255: 85%|########5 | 17/20 [00:23<00:04, 1.35s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000969 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.653975 valid's binary_logloss: 0.689905
Early stopping, best iteration is:
[63] train's binary_logloss: 0.665486 valid's binary_logloss: 0.689503
num_leaves, val_score: 0.688255: 90%|######### | 18/20 [00:24<00:02, 1.25s/it][I 2020-09-27 04:58:43,991] Trial 24 finished with value: 0.6895028538395134 and parameters: {'num_leaves': 47}. Best is trial 18 with value: 0.6882551844633495.
num_leaves, val_score: 0.688255: 90%|######### | 18/20 [00:24<00:02, 1.25s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009425 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688527 valid's binary_logloss: 0.690133
[200] train's binary_logloss: 0.686734 valid's binary_logloss: 0.689033
[300] train's binary_logloss: 0.685468 valid's binary_logloss: 0.688641
[400] train's binary_logloss: 0.684369 valid's binary_logloss: 0.688513
Early stopping, best iteration is:
[375] train's binary_logloss: 0.684633 valid's binary_logloss: 0.68848
num_leaves, val_score: 0.688255: 95%|#########5| 19/20 [00:26<00:01, 1.34s/it][I 2020-09-27 04:58:45,542] Trial 25 finished with value: 0.6884795876194431 and parameters: {'num_leaves': 3}. Best is trial 18 with value: 0.6882551844633495.
num_leaves, val_score: 0.688255: 95%|#########5| 19/20 [00:26<00:01, 1.34s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000942 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.634107 valid's binary_logloss: 0.691117
Early stopping, best iteration is:
[34] train's binary_logloss: 0.66756 valid's binary_logloss: 0.689318
num_leaves, val_score: 0.688255: 100%|##########| 20/20 [00:26<00:00, 1.21s/it][I 2020-09-27 04:58:46,434] Trial 26 finished with value: 0.6893183889206451 and parameters: {'num_leaves': 76}. Best is trial 18 with value: 0.6882551844633495.
num_leaves, val_score: 0.688255: 100%|##########| 20/20 [00:26<00:00, 1.35s/it]
bagging, val_score: 0.688255: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000882 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577
[200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588
[300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856
[400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313
[500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084
[600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043
[700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014
Early stopping, best iteration is:
[671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857
bagging, val_score: 0.687857: 10%|# | 1/10 [00:02<00:23, 2.59s/it][I 2020-09-27 04:58:49,036] Trial 27 finished with value: 0.6878571684251173 and parameters: {'bagging_fraction': 0.7285342265331846, 'bagging_freq': 5}. Best is trial 27 with value: 0.6878571684251173.
bagging, val_score: 0.687857: 10%|# | 1/10 [00:02<00:23, 2.59s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000976 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689824 valid's binary_logloss: 0.690686
[200] train's binary_logloss: 0.688555 valid's binary_logloss: 0.68967
[300] train's binary_logloss: 0.68776 valid's binary_logloss: 0.688991
[400] train's binary_logloss: 0.687201 valid's binary_logloss: 0.688421
[500] train's binary_logloss: 0.68677 valid's binary_logloss: 0.688248
[600] train's binary_logloss: 0.686428 valid's binary_logloss: 0.688112
[700] train's binary_logloss: 0.686154 valid's binary_logloss: 0.688132
Early stopping, best iteration is:
[670] train's binary_logloss: 0.686227 valid's binary_logloss: 0.688037
bagging, val_score: 0.687857: 20%|## | 2/10 [00:05<00:20, 2.62s/it][I 2020-09-27 04:58:51,737] Trial 28 finished with value: 0.6880373748785599 and parameters: {'bagging_fraction': 0.7377723938528014, 'bagging_freq': 5}. Best is trial 27 with value: 0.6878571684251173.
bagging, val_score: 0.687857: 20%|## | 2/10 [00:05<00:20, 2.62s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001053 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689798 valid's binary_logloss: 0.690553
[200] train's binary_logloss: 0.688509 valid's binary_logloss: 0.689611
[300] train's binary_logloss: 0.687723 valid's binary_logloss: 0.688922
[400] train's binary_logloss: 0.68716 valid's binary_logloss: 0.688433
[500] train's binary_logloss: 0.686722 valid's binary_logloss: 0.688211
[600] train's binary_logloss: 0.686375 valid's binary_logloss: 0.688204
Early stopping, best iteration is:
[555] train's binary_logloss: 0.686527 valid's binary_logloss: 0.688082
bagging, val_score: 0.687857: 30%|### | 3/10 [00:07<00:17, 2.49s/it][I 2020-09-27 04:58:53,937] Trial 29 finished with value: 0.6880818176819911 and parameters: {'bagging_fraction': 0.7152227689225483, 'bagging_freq': 5}. Best is trial 27 with value: 0.6878571684251173.
bagging, val_score: 0.687857: 30%|### | 3/10 [00:07<00:17, 2.49s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000973 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689833 valid's binary_logloss: 0.690598
[200] train's binary_logloss: 0.688544 valid's binary_logloss: 0.68966
[300] train's binary_logloss: 0.687749 valid's binary_logloss: 0.688918
[400] train's binary_logloss: 0.687185 valid's binary_logloss: 0.688393
[500] train's binary_logloss: 0.686741 valid's binary_logloss: 0.688141
[600] train's binary_logloss: 0.686404 valid's binary_logloss: 0.688102
Early stopping, best iteration is:
[555] train's binary_logloss: 0.686547 valid's binary_logloss: 0.688023
bagging, val_score: 0.687857: 40%|#### | 4/10 [00:09<00:14, 2.43s/it][I 2020-09-27 04:58:56,211] Trial 30 finished with value: 0.6880233595037442 and parameters: {'bagging_fraction': 0.7356006552369923, 'bagging_freq': 5}. Best is trial 27 with value: 0.6878571684251173.
bagging, val_score: 0.687857: 40%|#### | 4/10 [00:09<00:14, 2.43s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000978 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689826 valid's binary_logloss: 0.690712
[200] train's binary_logloss: 0.688569 valid's binary_logloss: 0.689662
[300] train's binary_logloss: 0.687746 valid's binary_logloss: 0.688891
[400] train's binary_logloss: 0.687195 valid's binary_logloss: 0.688435
[500] train's binary_logloss: 0.686757 valid's binary_logloss: 0.688266
[600] train's binary_logloss: 0.686422 valid's binary_logloss: 0.688159
Early stopping, best iteration is:
[552] train's binary_logloss: 0.686576 valid's binary_logloss: 0.688066
bagging, val_score: 0.687857: 50%|##### | 5/10 [00:12<00:11, 2.37s/it][I 2020-09-27 04:58:58,459] Trial 31 finished with value: 0.6880661932157487 and parameters: {'bagging_fraction': 0.7317912944017423, 'bagging_freq': 5}. Best is trial 27 with value: 0.6878571684251173.
bagging, val_score: 0.687857: 50%|##### | 5/10 [00:12<00:11, 2.37s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001775 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689822 valid's binary_logloss: 0.690641
[200] train's binary_logloss: 0.688546 valid's binary_logloss: 0.689616
[300] train's binary_logloss: 0.687753 valid's binary_logloss: 0.688948
[400] train's binary_logloss: 0.687196 valid's binary_logloss: 0.68841
[500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688198
[600] train's binary_logloss: 0.686419 valid's binary_logloss: 0.688131
Early stopping, best iteration is:
[555] train's binary_logloss: 0.686564 valid's binary_logloss: 0.688079
bagging, val_score: 0.687857: 60%|###### | 6/10 [00:14<00:09, 2.34s/it][I 2020-09-27 04:59:00,714] Trial 32 finished with value: 0.6880787095164863 and parameters: {'bagging_fraction': 0.7367657281985888, 'bagging_freq': 5}. Best is trial 27 with value: 0.6878571684251173.
bagging, val_score: 0.687857: 60%|###### | 6/10 [00:14<00:09, 2.34s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000997 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689824 valid's binary_logloss: 0.690686
[200] train's binary_logloss: 0.688555 valid's binary_logloss: 0.68967
[300] train's binary_logloss: 0.68776 valid's binary_logloss: 0.688991
[400] train's binary_logloss: 0.687201 valid's binary_logloss: 0.688421
[500] train's binary_logloss: 0.68677 valid's binary_logloss: 0.688248
[600] train's binary_logloss: 0.686428 valid's binary_logloss: 0.688112
[700] train's binary_logloss: 0.686154 valid's binary_logloss: 0.688132
Early stopping, best iteration is:
[670] train's binary_logloss: 0.686227 valid's binary_logloss: 0.688037
bagging, val_score: 0.687857: 70%|####### | 7/10 [00:16<00:07, 2.40s/it][I 2020-09-27 04:59:03,266] Trial 33 finished with value: 0.6880373748785599 and parameters: {'bagging_fraction': 0.7377865570355829, 'bagging_freq': 5}. Best is trial 27 with value: 0.6878571684251173.
bagging, val_score: 0.687857: 70%|####### | 7/10 [00:16<00:07, 2.40s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000994 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689832 valid's binary_logloss: 0.69068
[200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689603
[300] train's binary_logloss: 0.687772 valid's binary_logloss: 0.688883
[400] train's binary_logloss: 0.687221 valid's binary_logloss: 0.68836
[500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.688152
[600] train's binary_logloss: 0.686435 valid's binary_logloss: 0.688108
Early stopping, best iteration is:
[555] train's binary_logloss: 0.686576 valid's binary_logloss: 0.687995
bagging, val_score: 0.687857: 80%|######## | 8/10 [00:19<00:04, 2.35s/it][I 2020-09-27 04:59:05,479] Trial 34 finished with value: 0.6879951969996739 and parameters: {'bagging_fraction': 0.7515162282300969, 'bagging_freq': 5}. Best is trial 27 with value: 0.6878571684251173.
bagging, val_score: 0.687857: 80%|######## | 8/10 [00:19<00:04, 2.35s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001010 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689966 valid's binary_logloss: 0.690938
[200] train's binary_logloss: 0.688773 valid's binary_logloss: 0.689819
[300] train's binary_logloss: 0.688013 valid's binary_logloss: 0.689196
[400] train's binary_logloss: 0.687467 valid's binary_logloss: 0.688806
[500] train's binary_logloss: 0.687057 valid's binary_logloss: 0.688497
[600] train's binary_logloss: 0.686739 valid's binary_logloss: 0.688298
[700] train's binary_logloss: 0.686478 valid's binary_logloss: 0.688315
[800] train's binary_logloss: 0.686255 valid's binary_logloss: 0.688233
[900] train's binary_logloss: 0.686067 valid's binary_logloss: 0.688251
Early stopping, best iteration is:
[869] train's binary_logloss: 0.686123 valid's binary_logloss: 0.688191
bagging, val_score: 0.687857: 90%|######### | 9/10 [00:22<00:02, 2.68s/it][I 2020-09-27 04:59:08,934] Trial 35 finished with value: 0.6881912997605557 and parameters: {'bagging_fraction': 0.9421235215974794, 'bagging_freq': 4}. Best is trial 27 with value: 0.6878571684251173.
bagging, val_score: 0.687857: 90%|######### | 9/10 [00:22<00:02, 2.68s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001114 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689917 valid's binary_logloss: 0.690846
[200] train's binary_logloss: 0.688683 valid's binary_logloss: 0.689724
[300] train's binary_logloss: 0.68788 valid's binary_logloss: 0.689127
[400] train's binary_logloss: 0.687336 valid's binary_logloss: 0.6886
[500] train's binary_logloss: 0.686912 valid's binary_logloss: 0.688322
[600] train's binary_logloss: 0.686585 valid's binary_logloss: 0.688191
[700] train's binary_logloss: 0.686301 valid's binary_logloss: 0.688109
[800] train's binary_logloss: 0.686068 valid's binary_logloss: 0.688108
[900] train's binary_logloss: 0.685866 valid's binary_logloss: 0.688065
Early stopping, best iteration is:
[869] train's binary_logloss: 0.68593 valid's binary_logloss: 0.687939
bagging, val_score: 0.687857: 100%|##########| 10/10 [00:25<00:00, 2.86s/it][I 2020-09-27 04:59:12,229] Trial 36 finished with value: 0.6879386559102665 and parameters: {'bagging_fraction': 0.8372483508530265, 'bagging_freq': 7}. Best is trial 27 with value: 0.6878571684251173.
bagging, val_score: 0.687857: 100%|##########| 10/10 [00:25<00:00, 2.58s/it]
feature_fraction_stage2, val_score: 0.687857: 0%| | 0/6 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001013 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577
[200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588
[300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856
[400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313
[500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084
[600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043
[700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014
Early stopping, best iteration is:
[671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857
feature_fraction_stage2, val_score: 0.687857: 17%|#6 | 1/6 [00:02<00:12, 2.55s/it][I 2020-09-27 04:59:14,795] Trial 37 finished with value: 0.6878571684251175 and parameters: {'feature_fraction': 0.484}. Best is trial 37 with value: 0.6878571684251175.
feature_fraction_stage2, val_score: 0.687857: 17%|#6 | 1/6 [00:02<00:12, 2.55s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000890 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690634
[200] train's binary_logloss: 0.688568 valid's binary_logloss: 0.689696
[300] train's binary_logloss: 0.687761 valid's binary_logloss: 0.688942
[400] train's binary_logloss: 0.687217 valid's binary_logloss: 0.688449
[500] train's binary_logloss: 0.686776 valid's binary_logloss: 0.688296
[600] train's binary_logloss: 0.686427 valid's binary_logloss: 0.688198
Early stopping, best iteration is:
[552] train's binary_logloss: 0.686591 valid's binary_logloss: 0.688123
feature_fraction_stage2, val_score: 0.687857: 33%|###3 | 2/6 [00:04<00:09, 2.42s/it][I 2020-09-27 04:59:16,923] Trial 38 finished with value: 0.6881228060729746 and parameters: {'feature_fraction': 0.45199999999999996}. Best is trial 37 with value: 0.6878571684251175.
feature_fraction_stage2, val_score: 0.687857: 33%|###3 | 2/6 [00:04<00:09, 2.42s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000977 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577
[200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588
[300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856
[400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313
[500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084
[600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043
[700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014
Early stopping, best iteration is:
[671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857
feature_fraction_stage2, val_score: 0.687857: 50%|##### | 3/6 [00:07<00:07, 2.47s/it][I 2020-09-27 04:59:19,495] Trial 39 finished with value: 0.6878571684251175 and parameters: {'feature_fraction': 0.516}. Best is trial 37 with value: 0.6878571684251175.
feature_fraction_stage2, val_score: 0.687857: 50%|##### | 3/6 [00:07<00:07, 2.47s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000891 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689834 valid's binary_logloss: 0.690644
[200] train's binary_logloss: 0.688576 valid's binary_logloss: 0.689601
[300] train's binary_logloss: 0.687774 valid's binary_logloss: 0.688902
[400] train's binary_logloss: 0.687209 valid's binary_logloss: 0.688349
[500] train's binary_logloss: 0.686776 valid's binary_logloss: 0.688135
[600] train's binary_logloss: 0.68644 valid's binary_logloss: 0.688077
Early stopping, best iteration is:
[555] train's binary_logloss: 0.686591 valid's binary_logloss: 0.687962
feature_fraction_stage2, val_score: 0.687857: 67%|######6 | 4/6 [00:09<00:04, 2.37s/it][I 2020-09-27 04:59:21,646] Trial 40 finished with value: 0.6879623345018646 and parameters: {'feature_fraction': 0.42}. Best is trial 37 with value: 0.6878571684251175.
feature_fraction_stage2, val_score: 0.687857: 67%|######6 | 4/6 [00:09<00:04, 2.37s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001099 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689815 valid's binary_logloss: 0.690584
[200] train's binary_logloss: 0.688543 valid's binary_logloss: 0.689546
[300] train's binary_logloss: 0.687738 valid's binary_logloss: 0.688813
[400] train's binary_logloss: 0.687187 valid's binary_logloss: 0.688335
[500] train's binary_logloss: 0.686757 valid's binary_logloss: 0.688142
[600] train's binary_logloss: 0.686403 valid's binary_logloss: 0.688107
Early stopping, best iteration is:
[555] train's binary_logloss: 0.68656 valid's binary_logloss: 0.688002
feature_fraction_stage2, val_score: 0.687857: 83%|########3 | 5/6 [00:11<00:02, 2.33s/it][I 2020-09-27 04:59:23,892] Trial 41 finished with value: 0.6880017676081877 and parameters: {'feature_fraction': 0.58}. Best is trial 37 with value: 0.6878571684251175.
feature_fraction_stage2, val_score: 0.687857: 83%|########3 | 5/6 [00:11<00:02, 2.33s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001021 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689822 valid's binary_logloss: 0.690658
[200] train's binary_logloss: 0.688541 valid's binary_logloss: 0.689595
[300] train's binary_logloss: 0.687744 valid's binary_logloss: 0.688805
[400] train's binary_logloss: 0.687186 valid's binary_logloss: 0.688373
[500] train's binary_logloss: 0.686758 valid's binary_logloss: 0.688191
[600] train's binary_logloss: 0.68641 valid's binary_logloss: 0.688144
[700] train's binary_logloss: 0.686139 valid's binary_logloss: 0.688033
Early stopping, best iteration is:
[671] train's binary_logloss: 0.686211 valid's binary_logloss: 0.687908
feature_fraction_stage2, val_score: 0.687857: 100%|##########| 6/6 [00:14<00:00, 2.41s/it][I 2020-09-27 04:59:26,490] Trial 42 finished with value: 0.6879081347102528 and parameters: {'feature_fraction': 0.5479999999999999}. Best is trial 37 with value: 0.6878571684251175.
feature_fraction_stage2, val_score: 0.687857: 100%|##########| 6/6 [00:14<00:00, 2.38s/it]
regularization_factors, val_score: 0.687857: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000956 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689827 valid's binary_logloss: 0.690607
[200] train's binary_logloss: 0.688566 valid's binary_logloss: 0.689648
[300] train's binary_logloss: 0.687757 valid's binary_logloss: 0.688831
[400] train's binary_logloss: 0.687199 valid's binary_logloss: 0.688357
[500] train's binary_logloss: 0.686766 valid's binary_logloss: 0.688172
[600] train's binary_logloss: 0.686415 valid's binary_logloss: 0.688061
Early stopping, best iteration is:
[555] train's binary_logloss: 0.686571 valid's binary_logloss: 0.68802
regularization_factors, val_score: 0.687857: 5%|5 | 1/20 [00:02<00:41, 2.18s/it][I 2020-09-27 04:59:28,683] Trial 43 finished with value: 0.6880201782378496 and parameters: {'lambda_l1': 0.09717855595894546, 'lambda_l2': 2.599101232080336e-05}. Best is trial 43 with value: 0.6880201782378496.
regularization_factors, val_score: 0.687857: 5%|5 | 1/20 [00:02<00:41, 2.18s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001017 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689864 valid's binary_logloss: 0.690597
[200] train's binary_logloss: 0.688608 valid's binary_logloss: 0.68959
[300] train's binary_logloss: 0.687806 valid's binary_logloss: 0.688794
[400] train's binary_logloss: 0.687261 valid's binary_logloss: 0.688287
[500] train's binary_logloss: 0.686828 valid's binary_logloss: 0.688121
[600] train's binary_logloss: 0.686479 valid's binary_logloss: 0.68802
Early stopping, best iteration is:
[552] train's binary_logloss: 0.686646 valid's binary_logloss: 0.68794
regularization_factors, val_score: 0.687857: 10%|# | 2/20 [00:04<00:39, 2.19s/it][I 2020-09-27 04:59:30,884] Trial 44 finished with value: 0.6879401204038632 and parameters: {'lambda_l1': 2.27626769111772e-08, 'lambda_l2': 3.9029986997360395}. Best is trial 44 with value: 0.6879401204038632.
regularization_factors, val_score: 0.687857: 10%|# | 2/20 [00:04<00:39, 2.19s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000975 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577
[200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588
[300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856
[400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313
[500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084
[600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043
[700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014
Early stopping, best iteration is:
[671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857
regularization_factors, val_score: 0.687857: 15%|#5 | 3/20 [00:06<00:39, 2.30s/it][I 2020-09-27 04:59:33,437] Trial 45 finished with value: 0.6878571684254513 and parameters: {'lambda_l1': 6.262982988092863e-08, 'lambda_l2': 4.447887374533072e-08}. Best is trial 45 with value: 0.6878571684254513.
regularization_factors, val_score: 0.687857: 15%|#5 | 3/20 [00:06<00:39, 2.30s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000939 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577
[200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588
[300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856
[400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313
[500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084
[600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043
[700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014
Early stopping, best iteration is:
[671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857
regularization_factors, val_score: 0.687857: 20%|## | 4/20 [00:09<00:37, 2.37s/it][I 2020-09-27 04:59:35,984] Trial 46 finished with value: 0.6878571684251785 and parameters: {'lambda_l1': 1.0695178382125108e-08, 'lambda_l2': 1.4853383476880933e-08}. Best is trial 46 with value: 0.6878571684251785.
regularization_factors, val_score: 0.687857: 20%|## | 4/20 [00:09<00:37, 2.37s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000993 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577
[200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588
[300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856
[400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313
[500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084
[600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043
[700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014
Early stopping, best iteration is:
[671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857
regularization_factors, val_score: 0.687857: 25%|##5 | 5/20 [00:12<00:36, 2.43s/it][I 2020-09-27 04:59:38,551] Trial 47 finished with value: 0.6878571684252762 and parameters: {'lambda_l1': 1.8418017989423836e-08, 'lambda_l2': 1.1880874452318282e-08}. Best is trial 46 with value: 0.6878571684251785.
regularization_factors, val_score: 0.687857: 25%|##5 | 5/20 [00:12<00:36, 2.43s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000984 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577
[200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588
[300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856
[400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313
[500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084
[600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043
[700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014
Early stopping, best iteration is:
[671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857
regularization_factors, val_score: 0.687857: 30%|### | 6/20 [00:14<00:34, 2.48s/it][I 2020-09-27 04:59:41,131] Trial 48 finished with value: 0.687857168425301 and parameters: {'lambda_l1': 1.8864997222383136e-08, 'lambda_l2': 1.1856973045132024e-08}. Best is trial 46 with value: 0.6878571684251785.
regularization_factors, val_score: 0.687857: 30%|### | 6/20 [00:14<00:34, 2.48s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000984 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577
[200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588
[300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856
[400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313
[500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084
[600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043
[700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014
Early stopping, best iteration is:
[671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857
regularization_factors, val_score: 0.687857: 35%|###5 | 7/20 [00:17<00:32, 2.53s/it][I 2020-09-27 04:59:43,792] Trial 49 finished with value: 0.6878571685137861 and parameters: {'lambda_l1': 2.3008656218442958e-05, 'lambda_l2': 1.0076176570249505e-08}. Best is trial 46 with value: 0.6878571684251785.
regularization_factors, val_score: 0.687857: 35%|###5 | 7/20 [00:17<00:32, 2.53s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000984 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577
[200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588
[300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856
[400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313
[500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084
[600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043
[700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014
Early stopping, best iteration is:
[671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857
regularization_factors, val_score: 0.687857: 40%|#### | 8/20 [00:19<00:30, 2.54s/it][I 2020-09-27 04:59:46,353] Trial 50 finished with value: 0.6878571684048133 and parameters: {'lambda_l1': 1.0325498267148577e-05, 'lambda_l2': 9.965999410553949e-06}. Best is trial 50 with value: 0.6878571684048133.
regularization_factors, val_score: 0.687857: 40%|#### | 8/20 [00:19<00:30, 2.54s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001025 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577
[200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588
[300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856
[400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313
[500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084
[600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043
[700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014
Early stopping, best iteration is:
[671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857
regularization_factors, val_score: 0.687857: 45%|####5 | 9/20 [00:22<00:28, 2.55s/it][I 2020-09-27 04:59:48,931] Trial 51 finished with value: 0.6878571684428725 and parameters: {'lambda_l1': 1.7588444362019025e-05, 'lambda_l2': 8.259367220543736e-06}. Best is trial 50 with value: 0.6878571684048133.
regularization_factors, val_score: 0.687857: 45%|####5 | 9/20 [00:22<00:28, 2.55s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000992 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577
[200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588
[300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856
[400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313
[500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084
[600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043
[700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014
Early stopping, best iteration is:
[671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857
regularization_factors, val_score: 0.687857: 50%|##### | 10/20 [00:25<00:25, 2.57s/it][I 2020-09-27 04:59:51,534] Trial 52 finished with value: 0.6878571684244191 and parameters: {'lambda_l1': 5.589768247317659e-07, 'lambda_l2': 5.973648036426743e-07}. Best is trial 50 with value: 0.6878571684048133.
regularization_factors, val_score: 0.687857: 50%|##### | 10/20 [00:25<00:25, 2.57s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000986 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577
[200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588
[300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856
[400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313
[500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084
[600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043
[700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014
Early stopping, best iteration is:
[671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857
regularization_factors, val_score: 0.687857: 55%|#####5 | 11/20 [00:27<00:23, 2.56s/it][I 2020-09-27 04:59:54,086] Trial 53 finished with value: 0.6878571684026484 and parameters: {'lambda_l1': 1.4374237878755214e-06, 'lambda_l2': 4.663535296869216e-06}. Best is trial 53 with value: 0.6878571684026484.
regularization_factors, val_score: 0.687857: 55%|#####5 | 11/20 [00:27<00:23, 2.56s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000982 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577
[200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588
[300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856
[400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313
[500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084
[600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043
[700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014
Early stopping, best iteration is:
[671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857
regularization_factors, val_score: 0.687857: 60%|###### | 12/20 [00:30<00:20, 2.57s/it][I 2020-09-27 04:59:56,659] Trial 54 finished with value: 0.6878571684225362 and parameters: {'lambda_l1': 4.5280682776478785e-06, 'lambda_l2': 3.232830506287648e-06}. Best is trial 53 with value: 0.6878571684026484.
regularization_factors, val_score: 0.687857: 60%|###### | 12/20 [00:30<00:20, 2.57s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001020 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577
[200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588
[300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856
[400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313
[500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084
[600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043
[700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014
Early stopping, best iteration is:
[671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857
regularization_factors, val_score: 0.687857: 65%|######5 | 13/20 [00:32<00:17, 2.56s/it][I 2020-09-27 04:59:59,214] Trial 55 finished with value: 0.6878571684052713 and parameters: {'lambda_l1': 2.2260343593678247e-06, 'lambda_l2': 4.664732251084378e-06}. Best is trial 53 with value: 0.6878571684026484.
regularization_factors, val_score: 0.687857: 65%|######5 | 13/20 [00:32<00:17, 2.56s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001004 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577
[200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588
[300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856
[400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313
[500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084
[600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043
[700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014
Early stopping, best iteration is:
[671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857
regularization_factors, val_score: 0.687857: 70%|####### | 14/20 [00:35<00:15, 2.56s/it][I 2020-09-27 05:00:01,765] Trial 56 finished with value: 0.6878571684225463 and parameters: {'lambda_l1': 2.731664642116516e-06, 'lambda_l2': 2.242416999990266e-06}. Best is trial 53 with value: 0.6878571684026484.
regularization_factors, val_score: 0.687857: 70%|####### | 14/20 [00:35<00:15, 2.56s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000988 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577
[200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588
[300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856
[400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313
[500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084
[600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043
[700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014
Early stopping, best iteration is:
[671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857
regularization_factors, val_score: 0.687857: 75%|#######5 | 15/20 [00:37<00:12, 2.56s/it][I 2020-09-27 05:00:04,326] Trial 57 finished with value: 0.6878571684168346 and parameters: {'lambda_l1': 2.4032192398197603e-06, 'lambda_l2': 2.868804045681902e-06}. Best is trial 53 with value: 0.6878571684026484.
regularization_factors, val_score: 0.687857: 75%|#######5 | 15/20 [00:37<00:12, 2.56s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000974 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577
[200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588
[300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856
[400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313
[500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084
[600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043
[700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014
Early stopping, best iteration is:
[671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857
regularization_factors, val_score: 0.687857: 80%|######## | 16/20 [00:40<00:10, 2.56s/it][I 2020-09-27 05:00:06,891] Trial 58 finished with value: 0.6878571684147985 and parameters: {'lambda_l1': 3.739599166500671e-06, 'lambda_l2': 3.8681688367300176e-06}. Best is trial 53 with value: 0.6878571684026484.
regularization_factors, val_score: 0.687857: 80%|######## | 16/20 [00:40<00:10, 2.56s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000993 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577
[200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588
[300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856
[400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313
[500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084
[600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043
[700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014
Early stopping, best iteration is:
[671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857
regularization_factors, val_score: 0.687857: 85%|########5 | 17/20 [00:42<00:07, 2.56s/it][I 2020-09-27 05:00:09,442] Trial 59 finished with value: 0.6878571672506346 and parameters: {'lambda_l1': 2.703708966445098e-06, 'lambda_l2': 0.00019305898786142614}. Best is trial 59 with value: 0.6878571672506346.
regularization_factors, val_score: 0.687857: 85%|########5 | 17/20 [00:42<00:07, 2.56s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000960 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577
[200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588
[300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856
[400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313
[500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084
[600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043
[700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014
Early stopping, best iteration is:
[671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857
regularization_factors, val_score: 0.687857: 90%|######### | 18/20 [00:45<00:05, 2.56s/it][I 2020-09-27 05:00:11,996] Trial 60 finished with value: 0.6878571657826678 and parameters: {'lambda_l1': 9.271384341210084e-07, 'lambda_l2': 0.0004314431407377575}. Best is trial 60 with value: 0.6878571657826678.
regularization_factors, val_score: 0.687857: 90%|######### | 18/20 [00:45<00:05, 2.56s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000998 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577
[200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588
[300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856
[400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313
[500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084
[600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043
[700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014
Early stopping, best iteration is:
[671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857
regularization_factors, val_score: 0.687857: 95%|#########5| 19/20 [00:48<00:02, 2.56s/it][I 2020-09-27 05:00:14,562] Trial 61 finished with value: 0.6878571600694793 and parameters: {'lambda_l1': 8.018982307617414e-07, 'lambda_l2': 0.0013639030286262217}. Best is trial 61 with value: 0.6878571600694793.
regularization_factors, val_score: 0.687857: 95%|#########5| 19/20 [00:48<00:02, 2.56s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000973 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577
[200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588
[300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856
[400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313
[500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084
[600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043
[700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014
Early stopping, best iteration is:
[671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857
regularization_factors, val_score: 0.687857: 100%|##########| 20/20 [00:50<00:00, 2.55s/it][I 2020-09-27 05:00:17,099] Trial 62 finished with value: 0.6878571590590913 and parameters: {'lambda_l1': 5.337492288420711e-07, 'lambda_l2': 0.0015292766949980237}. Best is trial 62 with value: 0.6878571590590913.
regularization_factors, val_score: 0.687857: 100%|##########| 20/20 [00:50<00:00, 2.53s/it]
min_data_in_leaf, val_score: 0.687857: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001704 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577
[200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588
[300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856
[400] train's binary_logloss: 0.687194 valid's binary_logloss: 0.688269
[500] train's binary_logloss: 0.686757 valid's binary_logloss: 0.688121
[600] train's binary_logloss: 0.686414 valid's binary_logloss: 0.688068
Early stopping, best iteration is:
[551] train's binary_logloss: 0.686573 valid's binary_logloss: 0.687983
min_data_in_leaf, val_score: 0.687857: 20%|## | 1/5 [00:02<00:08, 2.20s/it][I 2020-09-27 05:00:19,310] Trial 63 finished with value: 0.6879830688856906 and parameters: {'min_child_samples': 25}. Best is trial 63 with value: 0.6879830688856906.
min_data_in_leaf, val_score: 0.687857: 20%|## | 1/5 [00:02<00:08, 2.20s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000979 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577
[200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588
[300] train's binary_logloss: 0.687742 valid's binary_logloss: 0.688876
[400] train's binary_logloss: 0.687187 valid's binary_logloss: 0.688301
[500] train's binary_logloss: 0.686743 valid's binary_logloss: 0.688115
[600] train's binary_logloss: 0.686394 valid's binary_logloss: 0.687995
Early stopping, best iteration is:
[555] train's binary_logloss: 0.686546 valid's binary_logloss: 0.68795
min_data_in_leaf, val_score: 0.687857: 40%|#### | 2/5 [00:06<00:08, 2.72s/it][I 2020-09-27 05:00:23,240] Trial 64 finished with value: 0.687949637690665 and parameters: {'min_child_samples': 10}. Best is trial 64 with value: 0.687949637690665.
min_data_in_leaf, val_score: 0.687857: 40%|#### | 2/5 [00:06<00:08, 2.72s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000925 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577
[200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588
[300] train's binary_logloss: 0.687742 valid's binary_logloss: 0.688876
[400] train's binary_logloss: 0.687187 valid's binary_logloss: 0.688301
[500] train's binary_logloss: 0.686741 valid's binary_logloss: 0.688135
[600] train's binary_logloss: 0.686393 valid's binary_logloss: 0.688019
Early stopping, best iteration is:
[553] train's binary_logloss: 0.68655 valid's binary_logloss: 0.687952
min_data_in_leaf, val_score: 0.687857: 60%|###### | 3/5 [00:08<00:05, 2.63s/it][I 2020-09-27 05:00:25,654] Trial 65 finished with value: 0.687952199445917 and parameters: {'min_child_samples': 5}. Best is trial 64 with value: 0.687949637690665.
min_data_in_leaf, val_score: 0.687857: 60%|###### | 3/5 [00:08<00:05, 2.63s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001028 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577
[200] train's binary_logloss: 0.688565 valid's binary_logloss: 0.689625
[300] train's binary_logloss: 0.687768 valid's binary_logloss: 0.688833
[400] train's binary_logloss: 0.687225 valid's binary_logloss: 0.688296
[500] train's binary_logloss: 0.686789 valid's binary_logloss: 0.688054
[600] train's binary_logloss: 0.686452 valid's binary_logloss: 0.687991
Early stopping, best iteration is:
[555] train's binary_logloss: 0.6866 valid's binary_logloss: 0.687924
min_data_in_leaf, val_score: 0.687857: 80%|######## | 4/5 [00:10<00:02, 2.52s/it][I 2020-09-27 05:00:27,934] Trial 66 finished with value: 0.6879239560726775 and parameters: {'min_child_samples': 50}. Best is trial 66 with value: 0.6879239560726775.
min_data_in_leaf, val_score: 0.687857: 80%|######## | 4/5 [00:10<00:02, 2.52s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000999 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062
[LightGBM] [Info] Start training from score -0.010062
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689877 valid's binary_logloss: 0.690659
[200] train's binary_logloss: 0.688626 valid's binary_logloss: 0.689631
[300] train's binary_logloss: 0.687853 valid's binary_logloss: 0.68889
[400] train's binary_logloss: 0.687343 valid's binary_logloss: 0.688334
[500] train's binary_logloss: 0.686947 valid's binary_logloss: 0.688179
[600] train's binary_logloss: 0.686609 valid's binary_logloss: 0.688099
Early stopping, best iteration is:
[555] train's binary_logloss: 0.686758 valid's binary_logloss: 0.688011
min_data_in_leaf, val_score: 0.687857: 100%|##########| 5/5 [00:13<00:00, 2.42s/it][I 2020-09-27 05:00:30,120] Trial 67 finished with value: 0.6880111488280128 and parameters: {'min_child_samples': 100}. Best is trial 66 with value: 0.6879239560726775.
min_data_in_leaf, val_score: 0.687857: 100%|##########| 5/5 [00:13<00:00, 2.60s/it]
Fold : 7
[I 2020-09-27 05:00:30,278] A new study created in memory with name: no-name-ad8f4155-35d8-4bc1-80b9-1a44d36999f7
feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000941 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665204 valid's binary_logloss: 0.691098
Early stopping, best iteration is:
[38] train's binary_logloss: 0.679318 valid's binary_logloss: 0.689859
feature_fraction, val_score: 0.689859: 14%|#4 | 1/7 [00:00<00:05, 1.15it/s][I 2020-09-27 05:00:31,155] Trial 0 finished with value: 0.6898590878905122 and parameters: {'feature_fraction': 0.5}. Best is trial 0 with value: 0.6898590878905122.
feature_fraction, val_score: 0.689859: 14%|#4 | 1/7 [00:00<00:05, 1.15it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000919 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.66602 valid's binary_logloss: 0.689895
Early stopping, best iteration is:
[60] train's binary_logloss: 0.674409 valid's binary_logloss: 0.689601
feature_fraction, val_score: 0.689601: 29%|##8 | 2/7 [00:01<00:04, 1.20it/s][I 2020-09-27 05:00:31,908] Trial 1 finished with value: 0.6896011115289385 and parameters: {'feature_fraction': 0.4}. Best is trial 1 with value: 0.6896011115289385.
feature_fraction, val_score: 0.689601: 29%|##8 | 2/7 [00:01<00:04, 1.20it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001642 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.662693 valid's binary_logloss: 0.690012
Early stopping, best iteration is:
[39] train's binary_logloss: 0.677876 valid's binary_logloss: 0.689256
feature_fraction, val_score: 0.689256: 43%|####2 | 3/7 [00:02<00:03, 1.23it/s][I 2020-09-27 05:00:32,673] Trial 2 finished with value: 0.6892558819835889 and parameters: {'feature_fraction': 1.0}. Best is trial 2 with value: 0.6892558819835889.
feature_fraction, val_score: 0.689256: 43%|####2 | 3/7 [00:02<00:03, 1.23it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001641 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663231 valid's binary_logloss: 0.690116
Early stopping, best iteration is:
[31] train's binary_logloss: 0.6803 valid's binary_logloss: 0.689625
feature_fraction, val_score: 0.689256: 57%|#####7 | 4/7 [00:03<00:02, 1.26it/s][I 2020-09-27 05:00:33,421] Trial 3 finished with value: 0.6896254736016139 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 2 with value: 0.6892558819835889.
feature_fraction, val_score: 0.689256: 57%|#####7 | 4/7 [00:03<00:02, 1.26it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012212 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664409 valid's binary_logloss: 0.68913
[200] train's binary_logloss: 0.644514 valid's binary_logloss: 0.690437
Early stopping, best iteration is:
[100] train's binary_logloss: 0.664409 valid's binary_logloss: 0.68913
feature_fraction, val_score: 0.689130: 71%|#######1 | 5/7 [00:04<00:01, 1.20it/s][I 2020-09-27 05:00:34,351] Trial 4 finished with value: 0.6891302495722187 and parameters: {'feature_fraction': 0.6}. Best is trial 4 with value: 0.6891302495722187.
feature_fraction, val_score: 0.689130: 71%|#######1 | 5/7 [00:04<00:01, 1.20it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016528 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664259 valid's binary_logloss: 0.689566
Early stopping, best iteration is:
[58] train's binary_logloss: 0.673563 valid's binary_logloss: 0.688962
feature_fraction, val_score: 0.688962: 86%|########5 | 6/7 [00:04<00:00, 1.21it/s][I 2020-09-27 05:00:35,160] Trial 5 finished with value: 0.6889623887000853 and parameters: {'feature_fraction': 0.7}. Best is trial 5 with value: 0.6889623887000853.
feature_fraction, val_score: 0.688962: 86%|########5 | 6/7 [00:04<00:00, 1.21it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001610 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663479 valid's binary_logloss: 0.689779
Early stopping, best iteration is:
[60] train's binary_logloss: 0.67284 valid's binary_logloss: 0.689141
feature_fraction, val_score: 0.688962: 100%|##########| 7/7 [00:05<00:00, 1.19it/s][I 2020-09-27 05:00:36,033] Trial 6 finished with value: 0.6891408826755168 and parameters: {'feature_fraction': 0.8}. Best is trial 5 with value: 0.6889623887000853.
feature_fraction, val_score: 0.688962: 100%|##########| 7/7 [00:05<00:00, 1.22it/s]
num_leaves, val_score: 0.688962: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002027 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.552919 valid's binary_logloss: 0.697247
Early stopping, best iteration is:
[23] train's binary_logloss: 0.648261 valid's binary_logloss: 0.691036
num_leaves, val_score: 0.688962: 5%|5 | 1/20 [00:01<00:27, 1.44s/it][I 2020-09-27 05:00:37,483] Trial 7 finished with value: 0.6910358844305945 and parameters: {'num_leaves': 211}. Best is trial 7 with value: 0.6910358844305945.
num_leaves, val_score: 0.688962: 5%|5 | 1/20 [00:01<00:27, 1.44s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007979 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.62836 valid's binary_logloss: 0.691669
Early stopping, best iteration is:
[27] train's binary_logloss: 0.669575 valid's binary_logloss: 0.689768
num_leaves, val_score: 0.688962: 10%|# | 2/20 [00:02<00:22, 1.25s/it][I 2020-09-27 05:00:38,292] Trial 8 finished with value: 0.6897684410879633 and parameters: {'num_leaves': 81}. Best is trial 8 with value: 0.6897684410879633.
num_leaves, val_score: 0.688962: 10%|# | 2/20 [00:02<00:22, 1.25s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001953 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.550486 valid's binary_logloss: 0.694876
Early stopping, best iteration is:
[27] train's binary_logloss: 0.641202 valid's binary_logloss: 0.689961
num_leaves, val_score: 0.688962: 15%|#5 | 3/20 [00:03<00:22, 1.31s/it][I 2020-09-27 05:00:39,748] Trial 9 finished with value: 0.6899614185647561 and parameters: {'num_leaves': 215}. Best is trial 8 with value: 0.6897684410879633.
num_leaves, val_score: 0.688962: 15%|#5 | 3/20 [00:03<00:22, 1.31s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012913 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685219 valid's binary_logloss: 0.689356
[200] train's binary_logloss: 0.681241 valid's binary_logloss: 0.689034
[300] train's binary_logloss: 0.677713 valid's binary_logloss: 0.689095
Early stopping, best iteration is:
[228] train's binary_logloss: 0.680209 valid's binary_logloss: 0.688922
num_leaves, val_score: 0.688922: 20%|## | 4/20 [00:04<00:20, 1.28s/it][I 2020-09-27 05:00:40,941] Trial 10 finished with value: 0.6889216574404491 and parameters: {'num_leaves': 6}. Best is trial 10 with value: 0.6889216574404491.
num_leaves, val_score: 0.688922: 20%|## | 4/20 [00:04<00:20, 1.28s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.006570 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687266 valid's binary_logloss: 0.68936
[200] train's binary_logloss: 0.684753 valid's binary_logloss: 0.688801
[300] train's binary_logloss: 0.682767 valid's binary_logloss: 0.688695
Early stopping, best iteration is:
[259] train's binary_logloss: 0.683544 valid's binary_logloss: 0.688542
num_leaves, val_score: 0.688542: 25%|##5 | 5/20 [00:06<00:19, 1.29s/it][I 2020-09-27 05:00:42,272] Trial 11 finished with value: 0.6885421835424413 and parameters: {'num_leaves': 4}. Best is trial 11 with value: 0.6885421835424413.
num_leaves, val_score: 0.688542: 25%|##5 | 5/20 [00:06<00:19, 1.29s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012743 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.683209 valid's binary_logloss: 0.688991
[200] train's binary_logloss: 0.677714 valid's binary_logloss: 0.689043
Early stopping, best iteration is:
[167] train's binary_logloss: 0.679404 valid's binary_logloss: 0.688853
num_leaves, val_score: 0.688542: 30%|### | 6/20 [00:07<00:17, 1.22s/it][I 2020-09-27 05:00:43,311] Trial 12 finished with value: 0.6888528740217242 and parameters: {'num_leaves': 8}. Best is trial 11 with value: 0.6885421835424413.
num_leaves, val_score: 0.688542: 30%|### | 6/20 [00:07<00:17, 1.22s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.022273 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687266 valid's binary_logloss: 0.68936
[200] train's binary_logloss: 0.684753 valid's binary_logloss: 0.688801
[300] train's binary_logloss: 0.682767 valid's binary_logloss: 0.688695
Early stopping, best iteration is:
[259] train's binary_logloss: 0.683544 valid's binary_logloss: 0.688542
num_leaves, val_score: 0.688542: 35%|###5 | 7/20 [00:08<00:16, 1.24s/it][I 2020-09-27 05:00:44,603] Trial 13 finished with value: 0.6885421835424413 and parameters: {'num_leaves': 4}. Best is trial 11 with value: 0.6885421835424413.
num_leaves, val_score: 0.688542: 35%|###5 | 7/20 [00:08<00:16, 1.24s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012970 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.636905 valid's binary_logloss: 0.691529
Early stopping, best iteration is:
[47] train's binary_logloss: 0.661435 valid's binary_logloss: 0.689547
num_leaves, val_score: 0.688542: 40%|#### | 8/20 [00:09<00:13, 1.14s/it][I 2020-09-27 05:00:45,518] Trial 14 finished with value: 0.6895472826915808 and parameters: {'num_leaves': 69}. Best is trial 11 with value: 0.6885421835424413.
num_leaves, val_score: 0.688542: 40%|#### | 8/20 [00:09<00:13, 1.14s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008973 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68623 valid's binary_logloss: 0.689472
[200] train's binary_logloss: 0.682968 valid's binary_logloss: 0.689396
Early stopping, best iteration is:
[166] train's binary_logloss: 0.683998 valid's binary_logloss: 0.689301
num_leaves, val_score: 0.688542: 45%|####5 | 9/20 [00:10<00:12, 1.09s/it][I 2020-09-27 05:00:46,499] Trial 15 finished with value: 0.6893008615554372 and parameters: {'num_leaves': 5}. Best is trial 11 with value: 0.6885421835424413.
num_leaves, val_score: 0.688542: 45%|####5 | 9/20 [00:10<00:12, 1.09s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008097 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.641155 valid's binary_logloss: 0.691233
Early stopping, best iteration is:
[44] train's binary_logloss: 0.665246 valid's binary_logloss: 0.689534
num_leaves, val_score: 0.688542: 50%|##### | 10/20 [00:11<00:10, 1.03s/it][I 2020-09-27 05:00:47,384] Trial 16 finished with value: 0.689534420595521 and parameters: {'num_leaves': 63}. Best is trial 11 with value: 0.6885421835424413.
num_leaves, val_score: 0.688542: 50%|##### | 10/20 [00:11<00:10, 1.03s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007854 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.594183 valid's binary_logloss: 0.693385
Early stopping, best iteration is:
[30] train's binary_logloss: 0.653722 valid's binary_logloss: 0.690977
num_leaves, val_score: 0.688542: 55%|#####5 | 11/20 [00:12<00:09, 1.04s/it][I 2020-09-27 05:00:48,441] Trial 17 finished with value: 0.6909774333414871 and parameters: {'num_leaves': 139}. Best is trial 11 with value: 0.6885421835424413.
num_leaves, val_score: 0.688542: 55%|#####5 | 11/20 [00:12<00:09, 1.04s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.014042 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.660928 valid's binary_logloss: 0.689724
Early stopping, best iteration is:
[65] train's binary_logloss: 0.669654 valid's binary_logloss: 0.689001
num_leaves, val_score: 0.688542: 60%|###### | 12/20 [00:13<00:07, 1.02it/s][I 2020-09-27 05:00:49,287] Trial 18 finished with value: 0.6890011375101475 and parameters: {'num_leaves': 35}. Best is trial 11 with value: 0.6885421835424413.
num_leaves, val_score: 0.688542: 60%|###### | 12/20 [00:13<00:07, 1.02it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007812 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.594516 valid's binary_logloss: 0.692963
Early stopping, best iteration is:
[16] train's binary_logloss: 0.670225 valid's binary_logloss: 0.691015
num_leaves, val_score: 0.688542: 65%|######5 | 13/20 [00:14<00:06, 1.03it/s][I 2020-09-27 05:00:50,248] Trial 19 finished with value: 0.6910149125364163 and parameters: {'num_leaves': 136}. Best is trial 11 with value: 0.6885421835424413.
num_leaves, val_score: 0.688542: 65%|######5 | 13/20 [00:14<00:06, 1.03it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008513 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.614802 valid's binary_logloss: 0.691487
Early stopping, best iteration is:
[50] train's binary_logloss: 0.64677 valid's binary_logloss: 0.690169
num_leaves, val_score: 0.688542: 70%|####### | 14/20 [00:15<00:05, 1.00it/s][I 2020-09-27 05:00:51,302] Trial 20 finished with value: 0.6901693018069993 and parameters: {'num_leaves': 103}. Best is trial 11 with value: 0.6885421835424413.
num_leaves, val_score: 0.688542: 70%|####### | 14/20 [00:15<00:05, 1.00it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001631 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68623 valid's binary_logloss: 0.689472
[200] train's binary_logloss: 0.682968 valid's binary_logloss: 0.689396
Early stopping, best iteration is:
[166] train's binary_logloss: 0.683998 valid's binary_logloss: 0.689301
num_leaves, val_score: 0.688542: 75%|#######5 | 15/20 [00:16<00:05, 1.00s/it][I 2020-09-27 05:00:52,310] Trial 21 finished with value: 0.6893008615554372 and parameters: {'num_leaves': 5}. Best is trial 11 with value: 0.6885421835424413.
num_leaves, val_score: 0.688542: 75%|#######5 | 15/20 [00:16<00:05, 1.00s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001749 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663209 valid's binary_logloss: 0.689372
Early stopping, best iteration is:
[56] train's binary_logloss: 0.673672 valid's binary_logloss: 0.688635
num_leaves, val_score: 0.688542: 80%|######## | 16/20 [00:17<00:03, 1.04it/s][I 2020-09-27 05:00:53,172] Trial 22 finished with value: 0.6886345495846673 and parameters: {'num_leaves': 32}. Best is trial 11 with value: 0.6885421835424413.
num_leaves, val_score: 0.688542: 80%|######## | 16/20 [00:17<00:03, 1.04it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008554 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.65661 valid's binary_logloss: 0.689805
Early stopping, best iteration is:
[44] train's binary_logloss: 0.673206 valid's binary_logloss: 0.689461
num_leaves, val_score: 0.688542: 85%|########5 | 17/20 [00:17<00:02, 1.10it/s][I 2020-09-27 05:00:53,968] Trial 23 finished with value: 0.6894608609879804 and parameters: {'num_leaves': 41}. Best is trial 11 with value: 0.6885421835424413.
num_leaves, val_score: 0.688542: 85%|########5 | 17/20 [00:17<00:02, 1.10it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012405 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.658212 valid's binary_logloss: 0.690617
Early stopping, best iteration is:
[33] train's binary_logloss: 0.677455 valid's binary_logloss: 0.689576
num_leaves, val_score: 0.688542: 90%|######### | 18/20 [00:18<00:01, 1.16it/s][I 2020-09-27 05:00:54,710] Trial 24 finished with value: 0.6895760118681855 and parameters: {'num_leaves': 39}. Best is trial 11 with value: 0.6885421835424413.
num_leaves, val_score: 0.688542: 90%|######### | 18/20 [00:18<00:01, 1.16it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008295 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.577324 valid's binary_logloss: 0.696256
Early stopping, best iteration is:
[31] train's binary_logloss: 0.645438 valid's binary_logloss: 0.691006
num_leaves, val_score: 0.688542: 95%|#########5| 19/20 [00:19<00:00, 1.03it/s][I 2020-09-27 05:00:55,935] Trial 25 finished with value: 0.6910055064236426 and parameters: {'num_leaves': 167}. Best is trial 11 with value: 0.6885421835424413.
num_leaves, val_score: 0.688542: 95%|#########5| 19/20 [00:19<00:00, 1.03it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012736 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667178 valid's binary_logloss: 0.689846
Early stopping, best iteration is:
[48] train's binary_logloss: 0.677808 valid's binary_logloss: 0.689292
num_leaves, val_score: 0.688542: 100%|##########| 20/20 [00:20<00:00, 1.11it/s][I 2020-09-27 05:00:56,664] Trial 26 finished with value: 0.6892917155718229 and parameters: {'num_leaves': 27}. Best is trial 11 with value: 0.6885421835424413.
num_leaves, val_score: 0.688542: 100%|##########| 20/20 [00:20<00:00, 1.03s/it]
bagging, val_score: 0.688542: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008116 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687216 valid's binary_logloss: 0.689193
[200] train's binary_logloss: 0.684606 valid's binary_logloss: 0.688599
[300] train's binary_logloss: 0.682514 valid's binary_logloss: 0.688495
Early stopping, best iteration is:
[243] train's binary_logloss: 0.683636 valid's binary_logloss: 0.688314
bagging, val_score: 0.688314: 10%|# | 1/10 [00:01<00:11, 1.33s/it][I 2020-09-27 05:00:58,010] Trial 27 finished with value: 0.6883135095365117 and parameters: {'bagging_fraction': 0.8631686996360303, 'bagging_freq': 3}. Best is trial 27 with value: 0.6883135095365117.
bagging, val_score: 0.688314: 10%|# | 1/10 [00:01<00:11, 1.33s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008381 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687216 valid's binary_logloss: 0.689219
[200] train's binary_logloss: 0.684738 valid's binary_logloss: 0.688577
[300] train's binary_logloss: 0.682586 valid's binary_logloss: 0.688474
[400] train's binary_logloss: 0.680625 valid's binary_logloss: 0.688466
Early stopping, best iteration is:
[388] train's binary_logloss: 0.680824 valid's binary_logloss: 0.688307
bagging, val_score: 0.688307: 20%|## | 2/10 [00:03<00:12, 1.50s/it][I 2020-09-27 05:00:59,908] Trial 28 finished with value: 0.6883067782443618 and parameters: {'bagging_fraction': 0.8747424021051551, 'bagging_freq': 3}. Best is trial 28 with value: 0.6883067782443618.
bagging, val_score: 0.688307: 20%|## | 2/10 [00:03<00:12, 1.50s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008348 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687228 valid's binary_logloss: 0.689193
[200] train's binary_logloss: 0.684624 valid's binary_logloss: 0.688721
[300] train's binary_logloss: 0.682522 valid's binary_logloss: 0.688655
[400] train's binary_logloss: 0.680572 valid's binary_logloss: 0.688787
Early stopping, best iteration is:
[313] train's binary_logloss: 0.682272 valid's binary_logloss: 0.688627
bagging, val_score: 0.688307: 30%|### | 3/10 [00:04<00:10, 1.54s/it][I 2020-09-27 05:01:01,550] Trial 29 finished with value: 0.6886270116658745 and parameters: {'bagging_fraction': 0.8767026255324745, 'bagging_freq': 3}. Best is trial 28 with value: 0.6883067782443618.
bagging, val_score: 0.688307: 30%|### | 3/10 [00:04<00:10, 1.54s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008608 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687185 valid's binary_logloss: 0.689143
[200] train's binary_logloss: 0.684623 valid's binary_logloss: 0.688512
[300] train's binary_logloss: 0.682521 valid's binary_logloss: 0.688305
Early stopping, best iteration is:
[299] train's binary_logloss: 0.682543 valid's binary_logloss: 0.688237
bagging, val_score: 0.688237: 40%|#### | 4/10 [00:06<00:09, 1.55s/it][I 2020-09-27 05:01:03,118] Trial 30 finished with value: 0.6882368432610401 and parameters: {'bagging_fraction': 0.8508850031981232, 'bagging_freq': 3}. Best is trial 30 with value: 0.6882368432610401.
bagging, val_score: 0.688237: 40%|#### | 4/10 [00:06<00:09, 1.55s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013843 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687182 valid's binary_logloss: 0.689138
[200] train's binary_logloss: 0.684621 valid's binary_logloss: 0.688476
[300] train's binary_logloss: 0.682518 valid's binary_logloss: 0.688405
[400] train's binary_logloss: 0.680515 valid's binary_logloss: 0.688508
Early stopping, best iteration is:
[307] train's binary_logloss: 0.682384 valid's binary_logloss: 0.688333
bagging, val_score: 0.688237: 50%|##### | 5/10 [00:08<00:07, 1.57s/it][I 2020-09-27 05:01:04,736] Trial 31 finished with value: 0.6883329240332864 and parameters: {'bagging_fraction': 0.8593267514380051, 'bagging_freq': 3}. Best is trial 30 with value: 0.6882368432610401.
bagging, val_score: 0.688237: 50%|##### | 5/10 [00:08<00:07, 1.57s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010252 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687215 valid's binary_logloss: 0.689327
[200] train's binary_logloss: 0.684669 valid's binary_logloss: 0.688643
[300] train's binary_logloss: 0.682522 valid's binary_logloss: 0.68885
[400] train's binary_logloss: 0.680567 valid's binary_logloss: 0.688669
Early stopping, best iteration is:
[353] train's binary_logloss: 0.681477 valid's binary_logloss: 0.68855
bagging, val_score: 0.688237: 60%|###### | 6/10 [00:09<00:06, 1.62s/it][I 2020-09-27 05:01:06,469] Trial 32 finished with value: 0.6885496852605667 and parameters: {'bagging_fraction': 0.8612598431625776, 'bagging_freq': 3}. Best is trial 30 with value: 0.6882368432610401.
bagging, val_score: 0.688237: 60%|###### | 6/10 [00:09<00:06, 1.62s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008437 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68717 valid's binary_logloss: 0.689286
[200] train's binary_logloss: 0.684607 valid's binary_logloss: 0.688529
[300] train's binary_logloss: 0.682498 valid's binary_logloss: 0.688434
Early stopping, best iteration is:
[262] train's binary_logloss: 0.683269 valid's binary_logloss: 0.688293
bagging, val_score: 0.688237: 70%|####### | 7/10 [00:11<00:04, 1.57s/it][I 2020-09-27 05:01:07,911] Trial 33 finished with value: 0.6882931441850038 and parameters: {'bagging_fraction': 0.8671876569591457, 'bagging_freq': 3}. Best is trial 30 with value: 0.6882368432610401.
bagging, val_score: 0.688237: 70%|####### | 7/10 [00:11<00:04, 1.57s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008157 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687191 valid's binary_logloss: 0.68922
[200] train's binary_logloss: 0.68462 valid's binary_logloss: 0.688602
Early stopping, best iteration is:
[186] train's binary_logloss: 0.684928 valid's binary_logloss: 0.688549
bagging, val_score: 0.688237: 80%|######## | 8/10 [00:12<00:02, 1.45s/it][I 2020-09-27 05:01:09,099] Trial 34 finished with value: 0.6885490341442905 and parameters: {'bagging_fraction': 0.8804920945207426, 'bagging_freq': 3}. Best is trial 30 with value: 0.6882368432610401.
bagging, val_score: 0.688237: 80%|######## | 8/10 [00:12<00:02, 1.45s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013640 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68703 valid's binary_logloss: 0.688776
[200] train's binary_logloss: 0.68453 valid's binary_logloss: 0.688164
[300] train's binary_logloss: 0.682427 valid's binary_logloss: 0.687911
[400] train's binary_logloss: 0.680592 valid's binary_logloss: 0.688176
Early stopping, best iteration is:
[347] train's binary_logloss: 0.681525 valid's binary_logloss: 0.687729
bagging, val_score: 0.687729: 90%|######### | 9/10 [00:14<00:01, 1.52s/it][I 2020-09-27 05:01:10,771] Trial 35 finished with value: 0.6877285343949435 and parameters: {'bagging_fraction': 0.6813927760377331, 'bagging_freq': 3}. Best is trial 35 with value: 0.6877285343949435.
bagging, val_score: 0.687729: 90%|######### | 9/10 [00:14<00:01, 1.52s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008939 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687118 valid's binary_logloss: 0.688984
[200] train's binary_logloss: 0.684488 valid's binary_logloss: 0.688477
[300] train's binary_logloss: 0.682515 valid's binary_logloss: 0.688435
[400] train's binary_logloss: 0.680676 valid's binary_logloss: 0.688358
[500] train's binary_logloss: 0.678963 valid's binary_logloss: 0.688512
Early stopping, best iteration is:
[456] train's binary_logloss: 0.679669 valid's binary_logloss: 0.688231
bagging, val_score: 0.687729: 100%|##########| 10/10 [00:15<00:00, 1.63s/it][I 2020-09-27 05:01:12,671] Trial 36 finished with value: 0.6882307476214391 and parameters: {'bagging_fraction': 0.6088933564022119, 'bagging_freq': 6}. Best is trial 35 with value: 0.6877285343949435.
bagging, val_score: 0.687729: 100%|##########| 10/10 [00:16<00:00, 1.60s/it]
feature_fraction_stage2, val_score: 0.687729: 0%| | 0/6 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001488 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687028 valid's binary_logloss: 0.688811
[200] train's binary_logloss: 0.684577 valid's binary_logloss: 0.68832
[300] train's binary_logloss: 0.682495 valid's binary_logloss: 0.688292
Early stopping, best iteration is:
[229] train's binary_logloss: 0.683926 valid's binary_logloss: 0.688061
feature_fraction_stage2, val_score: 0.687729: 17%|#6 | 1/6 [00:01<00:06, 1.31s/it][I 2020-09-27 05:01:13,995] Trial 37 finished with value: 0.6880605102479459 and parameters: {'feature_fraction': 0.7799999999999999}. Best is trial 37 with value: 0.6880605102479459.
feature_fraction_stage2, val_score: 0.687729: 17%|#6 | 1/6 [00:01<00:06, 1.31s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007740 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68703 valid's binary_logloss: 0.688776
[200] train's binary_logloss: 0.68453 valid's binary_logloss: 0.688164
[300] train's binary_logloss: 0.682427 valid's binary_logloss: 0.687911
[400] train's binary_logloss: 0.680592 valid's binary_logloss: 0.688176
Early stopping, best iteration is:
[347] train's binary_logloss: 0.681525 valid's binary_logloss: 0.687729
feature_fraction_stage2, val_score: 0.687729: 33%|###3 | 2/6 [00:03<00:05, 1.43s/it][I 2020-09-27 05:01:15,700] Trial 38 finished with value: 0.6877285343949435 and parameters: {'feature_fraction': 0.6839999999999999}. Best is trial 38 with value: 0.6877285343949435.
feature_fraction_stage2, val_score: 0.687729: 33%|###3 | 2/6 [00:03<00:05, 1.43s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009068 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68704 valid's binary_logloss: 0.688886
[200] train's binary_logloss: 0.684508 valid's binary_logloss: 0.688072
[300] train's binary_logloss: 0.682367 valid's binary_logloss: 0.688022
Early stopping, best iteration is:
[230] train's binary_logloss: 0.683826 valid's binary_logloss: 0.687901
feature_fraction_stage2, val_score: 0.687729: 50%|##### | 3/6 [00:04<00:04, 1.41s/it][I 2020-09-27 05:01:17,068] Trial 39 finished with value: 0.68790080420035 and parameters: {'feature_fraction': 0.716}. Best is trial 38 with value: 0.6877285343949435.
feature_fraction_stage2, val_score: 0.687729: 50%|##### | 3/6 [00:04<00:04, 1.41s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008155 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68704 valid's binary_logloss: 0.688886
[200] train's binary_logloss: 0.684508 valid's binary_logloss: 0.688072
[300] train's binary_logloss: 0.682367 valid's binary_logloss: 0.688022
Early stopping, best iteration is:
[230] train's binary_logloss: 0.683826 valid's binary_logloss: 0.687901
feature_fraction_stage2, val_score: 0.687729: 67%|######6 | 4/6 [00:05<00:02, 1.40s/it][I 2020-09-27 05:01:18,433] Trial 40 finished with value: 0.68790080420035 and parameters: {'feature_fraction': 0.748}. Best is trial 38 with value: 0.6877285343949435.
feature_fraction_stage2, val_score: 0.687729: 67%|######6 | 4/6 [00:05<00:02, 1.40s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010780 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737
[200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967
[300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694
Early stopping, best iteration is:
[231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553
feature_fraction_stage2, val_score: 0.687553: 83%|########3 | 5/6 [00:06<00:01, 1.35s/it][I 2020-09-27 05:01:19,668] Trial 41 finished with value: 0.6875532234561516 and parameters: {'feature_fraction': 0.652}. Best is trial 41 with value: 0.6875532234561516.
feature_fraction_stage2, val_score: 0.687553: 83%|########3 | 5/6 [00:06<00:01, 1.35s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.015763 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687015 valid's binary_logloss: 0.68889
[200] train's binary_logloss: 0.684543 valid's binary_logloss: 0.688419
[300] train's binary_logloss: 0.682467 valid's binary_logloss: 0.688343
Early stopping, best iteration is:
[234] train's binary_logloss: 0.683801 valid's binary_logloss: 0.688055
feature_fraction_stage2, val_score: 0.687553: 100%|##########| 6/6 [00:08<00:00, 1.32s/it][I 2020-09-27 05:01:20,912] Trial 42 finished with value: 0.6880548224298009 and parameters: {'feature_fraction': 0.62}. Best is trial 41 with value: 0.6875532234561516.
feature_fraction_stage2, val_score: 0.687553: 100%|##########| 6/6 [00:08<00:00, 1.37s/it]
regularization_factors, val_score: 0.687553: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009634 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737
[200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967
[300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694
Early stopping, best iteration is:
[231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553
regularization_factors, val_score: 0.687553: 5%|5 | 1/20 [00:01<00:24, 1.28s/it][I 2020-09-27 05:01:22,212] Trial 43 finished with value: 0.6875532234753943 and parameters: {'lambda_l1': 8.067143310213936e-07, 'lambda_l2': 1.7924277356165816e-07}. Best is trial 43 with value: 0.6875532234753943.
regularization_factors, val_score: 0.687553: 5%|5 | 1/20 [00:01<00:24, 1.28s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.005038 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737
[200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967
[300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694
Early stopping, best iteration is:
[231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553
regularization_factors, val_score: 0.687553: 10%|# | 2/20 [00:02<00:23, 1.30s/it][I 2020-09-27 05:01:23,564] Trial 44 finished with value: 0.6875532234639651 and parameters: {'lambda_l1': 2.9817465406512945e-07, 'lambda_l2': 7.314463336600674e-08}. Best is trial 44 with value: 0.6875532234639651.
regularization_factors, val_score: 0.687553: 10%|# | 2/20 [00:02<00:23, 1.30s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012896 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737
[200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967
[300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694
Early stopping, best iteration is:
[231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553
regularization_factors, val_score: 0.687553: 15%|#5 | 3/20 [00:03<00:21, 1.29s/it][I 2020-09-27 05:01:24,835] Trial 45 finished with value: 0.6875532234634629 and parameters: {'lambda_l1': 2.9860158446650325e-07, 'lambda_l2': 5.2722857836715734e-08}. Best is trial 45 with value: 0.6875532234634629.
regularization_factors, val_score: 0.687553: 15%|#5 | 3/20 [00:03<00:21, 1.29s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007909 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737
[200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967
[300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694
Early stopping, best iteration is:
[231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553
regularization_factors, val_score: 0.687553: 20%|## | 4/20 [00:05<00:20, 1.28s/it][I 2020-09-27 05:01:26,067] Trial 46 finished with value: 0.6875532234640681 and parameters: {'lambda_l1': 3.287625005492051e-07, 'lambda_l2': 3.053523237410299e-08}. Best is trial 45 with value: 0.6875532234634629.
regularization_factors, val_score: 0.687553: 20%|## | 4/20 [00:05<00:20, 1.28s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007843 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737
[200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967
[300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694
Early stopping, best iteration is:
[231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553
regularization_factors, val_score: 0.687553: 25%|##5 | 5/20 [00:06<00:19, 1.27s/it][I 2020-09-27 05:01:27,332] Trial 47 finished with value: 0.6875532234679212 and parameters: {'lambda_l1': 4.737421684219058e-07, 'lambda_l2': 3.2678236762325496e-08}. Best is trial 45 with value: 0.6875532234634629.
regularization_factors, val_score: 0.687553: 25%|##5 | 5/20 [00:06<00:19, 1.27s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012460 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737
[200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967
[300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694
Early stopping, best iteration is:
[231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553
regularization_factors, val_score: 0.687553: 30%|### | 6/20 [00:07<00:17, 1.27s/it][I 2020-09-27 05:01:28,588] Trial 48 finished with value: 0.6875532234603976 and parameters: {'lambda_l1': 1.8616761378629058e-07, 'lambda_l2': 1.614659705144355e-08}. Best is trial 48 with value: 0.6875532234603976.
regularization_factors, val_score: 0.687553: 30%|### | 6/20 [00:07<00:17, 1.27s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008626 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737
[200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967
[300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694
Early stopping, best iteration is:
[231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553
regularization_factors, val_score: 0.687553: 35%|###5 | 7/20 [00:08<00:16, 1.27s/it][I 2020-09-27 05:01:29,854] Trial 49 finished with value: 0.6875532234594542 and parameters: {'lambda_l1': 1.3735113285620687e-07, 'lambda_l2': 2.067788810166821e-08}. Best is trial 49 with value: 0.6875532234594542.
regularization_factors, val_score: 0.687553: 35%|###5 | 7/20 [00:08<00:16, 1.27s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012441 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737
[200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967
[300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694
Early stopping, best iteration is:
[231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553
regularization_factors, val_score: 0.687553: 40%|#### | 8/20 [00:10<00:15, 1.28s/it][I 2020-09-27 05:01:31,172] Trial 50 finished with value: 0.6875532234565727 and parameters: {'lambda_l1': 2.5154633280823624e-08, 'lambda_l2': 1.0923630752094391e-08}. Best is trial 50 with value: 0.6875532234565727.
regularization_factors, val_score: 0.687553: 40%|#### | 8/20 [00:10<00:15, 1.28s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002245 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737
[200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967
[300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694
Early stopping, best iteration is:
[231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553
regularization_factors, val_score: 0.687553: 45%|####5 | 9/20 [00:11<00:14, 1.29s/it][I 2020-09-27 05:01:32,467] Trial 51 finished with value: 0.6875532234566062 and parameters: {'lambda_l1': 3.0807637392676064e-08, 'lambda_l2': 1.4533537232191534e-08}. Best is trial 50 with value: 0.6875532234565727.
regularization_factors, val_score: 0.687553: 45%|####5 | 9/20 [00:11<00:14, 1.29s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008797 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737
[200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967
[300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694
Early stopping, best iteration is:
[231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553
regularization_factors, val_score: 0.687553: 50%|##### | 10/20 [00:12<00:12, 1.28s/it][I 2020-09-27 05:01:33,731] Trial 52 finished with value: 0.687553223456156 and parameters: {'lambda_l1': 1.0265554714929266e-08, 'lambda_l2': 1.0447547951042816e-08}. Best is trial 52 with value: 0.687553223456156.
regularization_factors, val_score: 0.687553: 50%|##### | 10/20 [00:12<00:12, 1.28s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001950 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737
[200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967
[300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694
Early stopping, best iteration is:
[231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553
regularization_factors, val_score: 0.687553: 55%|#####5 | 11/20 [00:14<00:11, 1.30s/it][I 2020-09-27 05:01:35,074] Trial 53 finished with value: 0.6875532234561593 and parameters: {'lambda_l1': 1.0392524686592004e-08, 'lambda_l2': 1.0663353352190143e-08}. Best is trial 52 with value: 0.687553223456156.
regularization_factors, val_score: 0.687553: 55%|#####5 | 11/20 [00:14<00:11, 1.30s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008108 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687064 valid's binary_logloss: 0.688739
[200] train's binary_logloss: 0.684486 valid's binary_logloss: 0.688006
[300] train's binary_logloss: 0.682501 valid's binary_logloss: 0.687947
Early stopping, best iteration is:
[231] train's binary_logloss: 0.683839 valid's binary_logloss: 0.687771
regularization_factors, val_score: 0.687553: 60%|###### | 12/20 [00:15<00:10, 1.29s/it][I 2020-09-27 05:01:36,332] Trial 54 finished with value: 0.6877705631041466 and parameters: {'lambda_l1': 2.1247372830919867e-08, 'lambda_l2': 0.06309677087509566}. Best is trial 52 with value: 0.687553223456156.
regularization_factors, val_score: 0.687553: 60%|###### | 12/20 [00:15<00:10, 1.29s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011180 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737
[200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967
[300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694
Early stopping, best iteration is:
[231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553
regularization_factors, val_score: 0.687553: 65%|######5 | 13/20 [00:16<00:08, 1.28s/it][I 2020-09-27 05:01:37,586] Trial 55 finished with value: 0.6875532234562124 and parameters: {'lambda_l1': 1.0739708350048745e-08, 'lambda_l2': 1.1070314813782383e-08}. Best is trial 52 with value: 0.687553223456156.
regularization_factors, val_score: 0.687553: 65%|######5 | 13/20 [00:16<00:08, 1.28s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008356 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687053 valid's binary_logloss: 0.68877
[200] train's binary_logloss: 0.684524 valid's binary_logloss: 0.688207
[300] train's binary_logloss: 0.682531 valid's binary_logloss: 0.687938
[400] train's binary_logloss: 0.68063 valid's binary_logloss: 0.688065
Early stopping, best iteration is:
[348] train's binary_logloss: 0.681599 valid's binary_logloss: 0.687655
regularization_factors, val_score: 0.687553: 70%|####### | 14/20 [00:18<00:08, 1.38s/it][I 2020-09-27 05:01:39,213] Trial 56 finished with value: 0.6876553612371981 and parameters: {'lambda_l1': 0.08111866585830418, 'lambda_l2': 7.066527702724071e-06}. Best is trial 52 with value: 0.687553223456156.
regularization_factors, val_score: 0.687553: 70%|####### | 14/20 [00:18<00:08, 1.38s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010624 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737
[200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967
[300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694
Early stopping, best iteration is:
[231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553
regularization_factors, val_score: 0.687553: 75%|#######5 | 15/20 [00:19<00:06, 1.35s/it][I 2020-09-27 05:01:40,473] Trial 57 finished with value: 0.6875532234561976 and parameters: {'lambda_l1': 1.0664783935126406e-08, 'lambda_l2': 1.0078447980735048e-08}. Best is trial 52 with value: 0.687553223456156.
regularization_factors, val_score: 0.687553: 75%|#######5 | 15/20 [00:19<00:06, 1.35s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007967 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737
[200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967
[300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694
Early stopping, best iteration is:
[231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553
regularization_factors, val_score: 0.687553: 80%|######## | 16/20 [00:20<00:05, 1.32s/it][I 2020-09-27 05:01:41,748] Trial 58 finished with value: 0.6875532234661581 and parameters: {'lambda_l1': 1.2662033446419092e-08, 'lambda_l2': 1.5025969082548396e-06}. Best is trial 52 with value: 0.687553223456156.
regularization_factors, val_score: 0.687553: 80%|######## | 16/20 [00:20<00:05, 1.32s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010265 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737
[200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967
[300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694
Early stopping, best iteration is:
[231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553
regularization_factors, val_score: 0.687553: 85%|########5 | 17/20 [00:22<00:03, 1.30s/it][I 2020-09-27 05:01:43,006] Trial 59 finished with value: 0.6875532234563088 and parameters: {'lambda_l1': 1.9318545060845645e-08, 'lambda_l2': 1.3003225260249327e-08}. Best is trial 52 with value: 0.687553223456156.
regularization_factors, val_score: 0.687553: 85%|########5 | 17/20 [00:22<00:03, 1.30s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008421 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737
[200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967
[300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694
Early stopping, best iteration is:
[231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553
regularization_factors, val_score: 0.687553: 90%|######### | 18/20 [00:23<00:02, 1.28s/it][I 2020-09-27 05:01:44,230] Trial 60 finished with value: 0.6875532237690561 and parameters: {'lambda_l1': 1.4255011647002522e-05, 'lambda_l2': 1.2469961061897813e-06}. Best is trial 52 with value: 0.687553223456156.
regularization_factors, val_score: 0.687553: 90%|######### | 18/20 [00:23<00:02, 1.28s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008041 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737
[200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967
[300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694
Early stopping, best iteration is:
[231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553
regularization_factors, val_score: 0.687553: 95%|#########5| 19/20 [00:24<00:01, 1.27s/it][I 2020-09-27 05:01:45,477] Trial 61 finished with value: 0.6875532234563587 and parameters: {'lambda_l1': 1.5764030850824877e-08, 'lambda_l2': 1.4202912414449929e-08}. Best is trial 52 with value: 0.687553223456156.
regularization_factors, val_score: 0.687553: 95%|#########5| 19/20 [00:24<00:01, 1.27s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007967 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737
[200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967
[300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694
Early stopping, best iteration is:
[231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553
regularization_factors, val_score: 0.687553: 100%|##########| 20/20 [00:25<00:00, 1.29s/it][I 2020-09-27 05:01:46,827] Trial 62 finished with value: 0.6875532234562169 and parameters: {'lambda_l1': 1.4154637690622022e-08, 'lambda_l2': 1.0757491754705414e-08}. Best is trial 52 with value: 0.687553223456156.
regularization_factors, val_score: 0.687553: 100%|##########| 20/20 [00:25<00:00, 1.30s/it]
min_data_in_leaf, val_score: 0.687553: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008157 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687147 valid's binary_logloss: 0.688925
[200] train's binary_logloss: 0.6848 valid's binary_logloss: 0.688284
[300] train's binary_logloss: 0.682838 valid's binary_logloss: 0.688287
Early stopping, best iteration is:
[229] train's binary_logloss: 0.684208 valid's binary_logloss: 0.688185
min_data_in_leaf, val_score: 0.687553: 20%|## | 1/5 [00:01<00:05, 1.47s/it][I 2020-09-27 05:01:48,314] Trial 63 finished with value: 0.688184603164505 and parameters: {'min_child_samples': 100}. Best is trial 63 with value: 0.688184603164505.
min_data_in_leaf, val_score: 0.687553: 20%|## | 1/5 [00:01<00:05, 1.47s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013040 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687067 valid's binary_logloss: 0.688767
[200] train's binary_logloss: 0.684549 valid's binary_logloss: 0.68803
[300] train's binary_logloss: 0.682536 valid's binary_logloss: 0.687853
Early stopping, best iteration is:
[228] train's binary_logloss: 0.683958 valid's binary_logloss: 0.687779
min_data_in_leaf, val_score: 0.687553: 40%|#### | 2/5 [00:02<00:04, 1.43s/it][I 2020-09-27 05:01:49,650] Trial 64 finished with value: 0.6877794897053392 and parameters: {'min_child_samples': 10}. Best is trial 64 with value: 0.6877794897053392.
min_data_in_leaf, val_score: 0.687553: 40%|#### | 2/5 [00:02<00:04, 1.43s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007873 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687067 valid's binary_logloss: 0.688767
[200] train's binary_logloss: 0.684546 valid's binary_logloss: 0.688205
[300] train's binary_logloss: 0.68251 valid's binary_logloss: 0.687847
[400] train's binary_logloss: 0.68055 valid's binary_logloss: 0.688135
Early stopping, best iteration is:
[347] train's binary_logloss: 0.681552 valid's binary_logloss: 0.687815
min_data_in_leaf, val_score: 0.687553: 60%|###### | 3/5 [00:04<00:02, 1.50s/it][I 2020-09-27 05:01:51,310] Trial 65 finished with value: 0.6878151679233732 and parameters: {'min_child_samples': 5}. Best is trial 64 with value: 0.6877794897053392.
min_data_in_leaf, val_score: 0.687553: 60%|###### | 3/5 [00:04<00:02, 1.50s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008336 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737
[200] train's binary_logloss: 0.684557 valid's binary_logloss: 0.68783
[300] train's binary_logloss: 0.682493 valid's binary_logloss: 0.687685
Early stopping, best iteration is:
[233] train's binary_logloss: 0.683814 valid's binary_logloss: 0.687524
min_data_in_leaf, val_score: 0.687524: 80%|######## | 4/5 [00:05<00:01, 1.42s/it][I 2020-09-27 05:01:52,557] Trial 66 finished with value: 0.6875238059643457 and parameters: {'min_child_samples': 25}. Best is trial 66 with value: 0.6875238059643457.
min_data_in_leaf, val_score: 0.687524: 80%|######## | 4/5 [00:05<00:01, 1.42s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008121 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245
[LightGBM] [Info] Start training from score -0.009245
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687098 valid's binary_logloss: 0.688869
[200] train's binary_logloss: 0.684669 valid's binary_logloss: 0.688242
[300] train's binary_logloss: 0.68261 valid's binary_logloss: 0.688041
[400] train's binary_logloss: 0.680742 valid's binary_logloss: 0.688478
Early stopping, best iteration is:
[302] train's binary_logloss: 0.682568 valid's binary_logloss: 0.688016
min_data_in_leaf, val_score: 0.687524: 100%|##########| 5/5 [00:07<00:00, 1.45s/it][I 2020-09-27 05:01:54,060] Trial 67 finished with value: 0.6880163054826545 and parameters: {'min_child_samples': 50}. Best is trial 66 with value: 0.6875238059643457.
min_data_in_leaf, val_score: 0.687524: 100%|##########| 5/5 [00:07<00:00, 1.44s/it]
Fold : 8
[I 2020-09-27 05:01:54,160] A new study created in memory with name: no-name-f83c045e-ae9d-45af-86f2-b4f88b2fc4c1
feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001841 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.6627 valid's binary_logloss: 0.6905
Early stopping, best iteration is:
[38] train's binary_logloss: 0.677925 valid's binary_logloss: 0.689727
feature_fraction, val_score: 0.689727: 14%|#4 | 1/7 [00:01<00:06, 1.03s/it][I 2020-09-27 05:01:55,205] Trial 0 finished with value: 0.6897267085894525 and parameters: {'feature_fraction': 1.0}. Best is trial 0 with value: 0.6897267085894525.
feature_fraction, val_score: 0.689727: 14%|#4 | 1/7 [00:01<00:06, 1.03s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000753 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.666345 valid's binary_logloss: 0.689793
Early stopping, best iteration is:
[40] train's binary_logloss: 0.679261 valid's binary_logloss: 0.689685
feature_fraction, val_score: 0.689685: 29%|##8 | 2/7 [00:01<00:04, 1.07it/s][I 2020-09-27 05:01:55,922] Trial 1 finished with value: 0.689684519676036 and parameters: {'feature_fraction': 0.4}. Best is trial 1 with value: 0.689684519676036.
feature_fraction, val_score: 0.689685: 29%|##8 | 2/7 [00:01<00:04, 1.07it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.033713 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664351 valid's binary_logloss: 0.689953
Early stopping, best iteration is:
[82] train's binary_logloss: 0.668343 valid's binary_logloss: 0.689942
feature_fraction, val_score: 0.689685: 43%|####2 | 3/7 [00:02<00:03, 1.02it/s][I 2020-09-27 05:01:56,995] Trial 2 finished with value: 0.6899415616595388 and parameters: {'feature_fraction': 0.6}. Best is trial 1 with value: 0.689684519676036.
feature_fraction, val_score: 0.689685: 43%|####2 | 3/7 [00:02<00:03, 1.02it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002095 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663723 valid's binary_logloss: 0.690154
Early stopping, best iteration is:
[47] train's binary_logloss: 0.676135 valid's binary_logloss: 0.689446
feature_fraction, val_score: 0.689446: 57%|#####7 | 4/7 [00:03<00:02, 1.08it/s][I 2020-09-27 05:01:57,810] Trial 3 finished with value: 0.6894460749095064 and parameters: {'feature_fraction': 0.7}. Best is trial 3 with value: 0.6894460749095064.
feature_fraction, val_score: 0.689446: 57%|#####7 | 4/7 [00:03<00:02, 1.08it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001658 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.662983 valid's binary_logloss: 0.690679
Early stopping, best iteration is:
[53] train's binary_logloss: 0.674243 valid's binary_logloss: 0.689539
feature_fraction, val_score: 0.689446: 71%|#######1 | 5/7 [00:04<00:01, 1.10it/s][I 2020-09-27 05:01:58,677] Trial 4 finished with value: 0.6895387621340154 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 3 with value: 0.6894460749095064.
feature_fraction, val_score: 0.689446: 71%|#######1 | 5/7 [00:04<00:01, 1.10it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008537 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663117 valid's binary_logloss: 0.689832
Early stopping, best iteration is:
[47] train's binary_logloss: 0.675766 valid's binary_logloss: 0.689575
feature_fraction, val_score: 0.689446: 86%|########5 | 6/7 [00:05<00:00, 1.10it/s][I 2020-09-27 05:01:59,592] Trial 5 finished with value: 0.6895746649335026 and parameters: {'feature_fraction': 0.8}. Best is trial 3 with value: 0.6894460749095064.
feature_fraction, val_score: 0.689446: 86%|########5 | 6/7 [00:05<00:00, 1.10it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000930 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665484 valid's binary_logloss: 0.690502
Early stopping, best iteration is:
[82] train's binary_logloss: 0.669205 valid's binary_logloss: 0.69007
feature_fraction, val_score: 0.689446: 100%|##########| 7/7 [00:06<00:00, 1.09it/s][I 2020-09-27 05:02:00,511] Trial 6 finished with value: 0.6900695949496937 and parameters: {'feature_fraction': 0.5}. Best is trial 3 with value: 0.6894460749095064.
feature_fraction, val_score: 0.689446: 100%|##########| 7/7 [00:06<00:00, 1.10it/s]
num_leaves, val_score: 0.689446: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008662 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.686038 valid's binary_logloss: 0.689757
[200] train's binary_logloss: 0.68281 valid's binary_logloss: 0.688899
[300] train's binary_logloss: 0.68005 valid's binary_logloss: 0.688585
[400] train's binary_logloss: 0.677584 valid's binary_logloss: 0.688509
Early stopping, best iteration is:
[336] train's binary_logloss: 0.679139 valid's binary_logloss: 0.688453
num_leaves, val_score: 0.688453: 5%|5 | 1/20 [00:01<00:30, 1.59s/it][I 2020-09-27 05:02:02,118] Trial 7 finished with value: 0.6884528114970678 and parameters: {'num_leaves': 5}. Best is trial 7 with value: 0.6884528114970678.
num_leaves, val_score: 0.688453: 5%|5 | 1/20 [00:01<00:30, 1.59s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012715 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.593349 valid's binary_logloss: 0.694762
Early stopping, best iteration is:
[28] train's binary_logloss: 0.655278 valid's binary_logloss: 0.690329
num_leaves, val_score: 0.688453: 10%|# | 2/20 [00:02<00:25, 1.43s/it][I 2020-09-27 05:02:03,178] Trial 8 finished with value: 0.690328689997322 and parameters: {'num_leaves': 139}. Best is trial 7 with value: 0.6884528114970678.
num_leaves, val_score: 0.688453: 10%|# | 2/20 [00:02<00:25, 1.43s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012613 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.545663 valid's binary_logloss: 0.696241
Early stopping, best iteration is:
[18] train's binary_logloss: 0.654494 valid's binary_logloss: 0.690387
num_leaves, val_score: 0.688453: 15%|#5 | 3/20 [00:03<00:23, 1.37s/it][I 2020-09-27 05:02:04,416] Trial 9 finished with value: 0.6903874648728597 and parameters: {'num_leaves': 227}. Best is trial 7 with value: 0.6884528114970678.
num_leaves, val_score: 0.688453: 15%|#5 | 3/20 [00:03<00:23, 1.37s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012246 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689978 valid's binary_logloss: 0.691198
[200] train's binary_logloss: 0.688797 valid's binary_logloss: 0.69055
[300] train's binary_logloss: 0.688072 valid's binary_logloss: 0.690072
[400] train's binary_logloss: 0.687556 valid's binary_logloss: 0.689833
[500] train's binary_logloss: 0.687169 valid's binary_logloss: 0.689688
[600] train's binary_logloss: 0.686868 valid's binary_logloss: 0.689597
[700] train's binary_logloss: 0.686624 valid's binary_logloss: 0.689542
[800] train's binary_logloss: 0.68642 valid's binary_logloss: 0.689474
[900] train's binary_logloss: 0.686246 valid's binary_logloss: 0.689462
[1000] train's binary_logloss: 0.686095 valid's binary_logloss: 0.689411
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.686095 valid's binary_logloss: 0.689411
num_leaves, val_score: 0.688453: 20%|## | 4/20 [00:06<00:30, 1.89s/it][I 2020-09-27 05:02:07,521] Trial 10 finished with value: 0.6894108839720557 and parameters: {'num_leaves': 2}. Best is trial 7 with value: 0.6884528114970678.
num_leaves, val_score: 0.688453: 20%|## | 4/20 [00:06<00:30, 1.89s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008181 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.684097 valid's binary_logloss: 0.689543
[200] train's binary_logloss: 0.679346 valid's binary_logloss: 0.689259
[300] train's binary_logloss: 0.675305 valid's binary_logloss: 0.68922
[400] train's binary_logloss: 0.671448 valid's binary_logloss: 0.689174
Early stopping, best iteration is:
[385] train's binary_logloss: 0.672024 valid's binary_logloss: 0.689088
num_leaves, val_score: 0.688453: 25%|##5 | 5/20 [00:08<00:28, 1.88s/it][I 2020-09-27 05:02:09,364] Trial 11 finished with value: 0.6890882300993826 and parameters: {'num_leaves': 7}. Best is trial 7 with value: 0.6884528114970678.
num_leaves, val_score: 0.688453: 25%|##5 | 5/20 [00:08<00:28, 1.88s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009553 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.684097 valid's binary_logloss: 0.689543
[200] train's binary_logloss: 0.679346 valid's binary_logloss: 0.689259
[300] train's binary_logloss: 0.675305 valid's binary_logloss: 0.68922
[400] train's binary_logloss: 0.671448 valid's binary_logloss: 0.689174
Early stopping, best iteration is:
[385] train's binary_logloss: 0.672024 valid's binary_logloss: 0.689088
num_leaves, val_score: 0.688453: 30%|### | 6/20 [00:10<00:25, 1.85s/it][I 2020-09-27 05:02:11,153] Trial 12 finished with value: 0.6890882300993826 and parameters: {'num_leaves': 7}. Best is trial 7 with value: 0.6884528114970678.
num_leaves, val_score: 0.688453: 30%|### | 6/20 [00:10<00:25, 1.85s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008338 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.652737 valid's binary_logloss: 0.690159
Early stopping, best iteration is:
[30] train's binary_logloss: 0.676594 valid's binary_logloss: 0.689592
num_leaves, val_score: 0.688453: 35%|###5 | 7/20 [00:11<00:19, 1.52s/it][I 2020-09-27 05:02:11,891] Trial 13 finished with value: 0.6895920137463518 and parameters: {'num_leaves': 46}. Best is trial 7 with value: 0.6884528114970678.
num_leaves, val_score: 0.688453: 35%|###5 | 7/20 [00:11<00:19, 1.52s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008791 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.630115 valid's binary_logloss: 0.692022
Early stopping, best iteration is:
[25] train's binary_logloss: 0.671325 valid's binary_logloss: 0.690619
num_leaves, val_score: 0.688453: 40%|#### | 8/20 [00:12<00:15, 1.33s/it][I 2020-09-27 05:02:12,774] Trial 14 finished with value: 0.690619286117565 and parameters: {'num_leaves': 79}. Best is trial 7 with value: 0.6884528114970678.
num_leaves, val_score: 0.688453: 40%|#### | 8/20 [00:12<00:15, 1.33s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008004 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.614161 valid's binary_logloss: 0.692813
Early stopping, best iteration is:
[31] train's binary_logloss: 0.661053 valid's binary_logloss: 0.690497
num_leaves, val_score: 0.688453: 45%|####5 | 9/20 [00:13<00:13, 1.21s/it][I 2020-09-27 05:02:13,713] Trial 15 finished with value: 0.6904972906779976 and parameters: {'num_leaves': 104}. Best is trial 7 with value: 0.6884528114970678.
num_leaves, val_score: 0.688453: 45%|####5 | 9/20 [00:13<00:13, 1.21s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008054 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685119 valid's binary_logloss: 0.68988
[200] train's binary_logloss: 0.681155 valid's binary_logloss: 0.689507
Early stopping, best iteration is:
[189] train's binary_logloss: 0.681535 valid's binary_logloss: 0.689459
num_leaves, val_score: 0.688453: 50%|##### | 10/20 [00:14<00:11, 1.17s/it][I 2020-09-27 05:02:14,772] Trial 16 finished with value: 0.6894592704386628 and parameters: {'num_leaves': 6}. Best is trial 7 with value: 0.6884528114970678.
num_leaves, val_score: 0.688453: 50%|##### | 10/20 [00:14<00:11, 1.17s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009175 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.568944 valid's binary_logloss: 0.695405
Early stopping, best iteration is:
[15] train's binary_logloss: 0.665473 valid's binary_logloss: 0.690687
num_leaves, val_score: 0.688453: 55%|#####5 | 11/20 [00:15<00:10, 1.13s/it][I 2020-09-27 05:02:15,813] Trial 17 finished with value: 0.6906872291686416 and parameters: {'num_leaves': 182}. Best is trial 7 with value: 0.6884528114970678.
num_leaves, val_score: 0.688453: 55%|#####5 | 11/20 [00:15<00:10, 1.13s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001779 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.649856 valid's binary_logloss: 0.690204
Early stopping, best iteration is:
[48] train's binary_logloss: 0.667974 valid's binary_logloss: 0.689828
num_leaves, val_score: 0.688453: 60%|###### | 12/20 [00:16<00:08, 1.06s/it][I 2020-09-27 05:02:16,731] Trial 18 finished with value: 0.6898284619368371 and parameters: {'num_leaves': 50}. Best is trial 7 with value: 0.6884528114970678.
num_leaves, val_score: 0.688453: 60%|###### | 12/20 [00:16<00:08, 1.06s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009546 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.652737 valid's binary_logloss: 0.690159
Early stopping, best iteration is:
[30] train's binary_logloss: 0.676594 valid's binary_logloss: 0.689592
num_leaves, val_score: 0.688453: 65%|######5 | 13/20 [00:16<00:06, 1.03it/s][I 2020-09-27 05:02:17,476] Trial 19 finished with value: 0.6895920137463518 and parameters: {'num_leaves': 46}. Best is trial 7 with value: 0.6884528114970678.
num_leaves, val_score: 0.688453: 65%|######5 | 13/20 [00:16<00:06, 1.03it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011684 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.688384 valid's binary_logloss: 0.6906
[200] train's binary_logloss: 0.686574 valid's binary_logloss: 0.689853
[300] train's binary_logloss: 0.68527 valid's binary_logloss: 0.689427
[400] train's binary_logloss: 0.684182 valid's binary_logloss: 0.689235
[500] train's binary_logloss: 0.683198 valid's binary_logloss: 0.689083
[600] train's binary_logloss: 0.6823 valid's binary_logloss: 0.688973
[700] train's binary_logloss: 0.681468 valid's binary_logloss: 0.688875
[800] train's binary_logloss: 0.680639 valid's binary_logloss: 0.688913
Early stopping, best iteration is:
[703] train's binary_logloss: 0.681443 valid's binary_logloss: 0.68886
num_leaves, val_score: 0.688453: 70%|####### | 14/20 [00:19<00:08, 1.48s/it][I 2020-09-27 05:02:20,147] Trial 20 finished with value: 0.6888601909760304 and parameters: {'num_leaves': 3}. Best is trial 7 with value: 0.6884528114970678.
num_leaves, val_score: 0.688453: 70%|####### | 14/20 [00:19<00:08, 1.48s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008648 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.676993 valid's binary_logloss: 0.689308
Early stopping, best iteration is:
[90] train's binary_logloss: 0.678106 valid's binary_logloss: 0.689282
num_leaves, val_score: 0.688453: 75%|#######5 | 15/20 [00:20<00:06, 1.28s/it][I 2020-09-27 05:02:20,976] Trial 21 finished with value: 0.6892821753767159 and parameters: {'num_leaves': 15}. Best is trial 7 with value: 0.6884528114970678.
num_leaves, val_score: 0.688453: 75%|#######5 | 15/20 [00:20<00:06, 1.28s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008704 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.689978 valid's binary_logloss: 0.691198
[200] train's binary_logloss: 0.688797 valid's binary_logloss: 0.69055
[300] train's binary_logloss: 0.688072 valid's binary_logloss: 0.690072
[400] train's binary_logloss: 0.687556 valid's binary_logloss: 0.689833
[500] train's binary_logloss: 0.687169 valid's binary_logloss: 0.689688
[600] train's binary_logloss: 0.686868 valid's binary_logloss: 0.689597
[700] train's binary_logloss: 0.686624 valid's binary_logloss: 0.689542
[800] train's binary_logloss: 0.68642 valid's binary_logloss: 0.689474
[900] train's binary_logloss: 0.686246 valid's binary_logloss: 0.689462
[1000] train's binary_logloss: 0.686095 valid's binary_logloss: 0.689411
Did not meet early stopping. Best iteration is:
[1000] train's binary_logloss: 0.686095 valid's binary_logloss: 0.689411
num_leaves, val_score: 0.688453: 80%|######## | 16/20 [00:25<00:09, 2.43s/it][I 2020-09-27 05:02:26,077] Trial 22 finished with value: 0.6894108839720557 and parameters: {'num_leaves': 2}. Best is trial 7 with value: 0.6884528114970678.
num_leaves, val_score: 0.688453: 80%|######## | 16/20 [00:25<00:09, 2.43s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016684 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.65797 valid's binary_logloss: 0.690213
Early stopping, best iteration is:
[38] train's binary_logloss: 0.675789 valid's binary_logloss: 0.689489
num_leaves, val_score: 0.688453: 85%|########5 | 17/20 [00:26<00:05, 1.96s/it][I 2020-09-27 05:02:26,941] Trial 23 finished with value: 0.6894891049398579 and parameters: {'num_leaves': 39}. Best is trial 7 with value: 0.6884528114970678.
num_leaves, val_score: 0.688453: 85%|########5 | 17/20 [00:26<00:05, 1.96s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011681 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.632096 valid's binary_logloss: 0.691354
Early stopping, best iteration is:
[28] train's binary_logloss: 0.670154 valid's binary_logloss: 0.690155
num_leaves, val_score: 0.688453: 90%|######### | 18/20 [00:27<00:03, 1.64s/it][I 2020-09-27 05:02:27,818] Trial 24 finished with value: 0.6901553410011205 and parameters: {'num_leaves': 75}. Best is trial 7 with value: 0.6884528114970678.
num_leaves, val_score: 0.688453: 90%|######### | 18/20 [00:27<00:03, 1.64s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012016 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.669171 valid's binary_logloss: 0.689591
Early stopping, best iteration is:
[78] train's binary_logloss: 0.67319 valid's binary_logloss: 0.68921
num_leaves, val_score: 0.688453: 95%|#########5| 19/20 [00:28<00:01, 1.42s/it][I 2020-09-27 05:02:28,738] Trial 25 finished with value: 0.6892103907051438 and parameters: {'num_leaves': 24}. Best is trial 7 with value: 0.6884528114970678.
num_leaves, val_score: 0.688453: 95%|#########5| 19/20 [00:28<00:01, 1.42s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.018748 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.631577 valid's binary_logloss: 0.692299
Early stopping, best iteration is:
[26] train's binary_logloss: 0.671068 valid's binary_logloss: 0.690199
num_leaves, val_score: 0.688453: 100%|##########| 20/20 [00:29<00:00, 1.26s/it][I 2020-09-27 05:02:29,625] Trial 26 finished with value: 0.6901985363739374 and parameters: {'num_leaves': 77}. Best is trial 7 with value: 0.6884528114970678.
num_leaves, val_score: 0.688453: 100%|##########| 20/20 [00:29<00:00, 1.46s/it]
bagging, val_score: 0.688453: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001613 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.686006 valid's binary_logloss: 0.689628
[200] train's binary_logloss: 0.682779 valid's binary_logloss: 0.688355
[300] train's binary_logloss: 0.68005 valid's binary_logloss: 0.688419
Early stopping, best iteration is:
[219] train's binary_logloss: 0.682239 valid's binary_logloss: 0.688258
bagging, val_score: 0.688258: 10%|# | 1/10 [00:01<00:11, 1.23s/it][I 2020-09-27 05:02:30,868] Trial 27 finished with value: 0.6882582742878182 and parameters: {'bagging_fraction': 0.615338737094182, 'bagging_freq': 7}. Best is trial 27 with value: 0.6882582742878182.
bagging, val_score: 0.688258: 10%|# | 1/10 [00:01<00:11, 1.23s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008365 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68604 valid's binary_logloss: 0.689719
[200] train's binary_logloss: 0.682712 valid's binary_logloss: 0.68842
Early stopping, best iteration is:
[194] train's binary_logloss: 0.682897 valid's binary_logloss: 0.688374
bagging, val_score: 0.688258: 20%|## | 2/10 [00:02<00:09, 1.19s/it][I 2020-09-27 05:02:31,952] Trial 28 finished with value: 0.6883743964950245 and parameters: {'bagging_fraction': 0.6026436236346611, 'bagging_freq': 7}. Best is trial 27 with value: 0.6882582742878182.
bagging, val_score: 0.688258: 20%|## | 2/10 [00:02<00:09, 1.19s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008715 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.686015 valid's binary_logloss: 0.689456
[200] train's binary_logloss: 0.682706 valid's binary_logloss: 0.688437
[300] train's binary_logloss: 0.67996 valid's binary_logloss: 0.688695
Early stopping, best iteration is:
[223] train's binary_logloss: 0.682071 valid's binary_logloss: 0.68839
bagging, val_score: 0.688258: 30%|### | 3/10 [00:03<00:08, 1.19s/it][I 2020-09-27 05:02:33,149] Trial 29 finished with value: 0.6883895614361045 and parameters: {'bagging_fraction': 0.6029377567256127, 'bagging_freq': 7}. Best is trial 27 with value: 0.6882582742878182.
bagging, val_score: 0.688258: 30%|### | 3/10 [00:03<00:08, 1.19s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008131 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685948 valid's binary_logloss: 0.689692
[200] train's binary_logloss: 0.68277 valid's binary_logloss: 0.688533
[300] train's binary_logloss: 0.679958 valid's binary_logloss: 0.688725
Early stopping, best iteration is:
[274] train's binary_logloss: 0.680722 valid's binary_logloss: 0.688429
bagging, val_score: 0.688258: 40%|#### | 4/10 [00:04<00:07, 1.22s/it][I 2020-09-27 05:02:34,457] Trial 30 finished with value: 0.6884290873607732 and parameters: {'bagging_fraction': 0.5938506890500748, 'bagging_freq': 7}. Best is trial 27 with value: 0.6882582742878182.
bagging, val_score: 0.688258: 40%|#### | 4/10 [00:04<00:07, 1.22s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013062 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685972 valid's binary_logloss: 0.689581
[200] train's binary_logloss: 0.682787 valid's binary_logloss: 0.688637
Early stopping, best iteration is:
[196] train's binary_logloss: 0.682904 valid's binary_logloss: 0.688622
bagging, val_score: 0.688258: 50%|##### | 5/10 [00:05<00:05, 1.18s/it][I 2020-09-27 05:02:35,525] Trial 31 finished with value: 0.6886216049092391 and parameters: {'bagging_fraction': 0.5957530118583011, 'bagging_freq': 7}. Best is trial 27 with value: 0.6882582742878182.
bagging, val_score: 0.688258: 50%|##### | 5/10 [00:05<00:05, 1.18s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008942 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.686042 valid's binary_logloss: 0.68983
[200] train's binary_logloss: 0.682839 valid's binary_logloss: 0.688484
Early stopping, best iteration is:
[196] train's binary_logloss: 0.682969 valid's binary_logloss: 0.688429
bagging, val_score: 0.688258: 60%|###### | 6/10 [00:06<00:04, 1.15s/it][I 2020-09-27 05:02:36,622] Trial 32 finished with value: 0.6884291730078321 and parameters: {'bagging_fraction': 0.601131517375276, 'bagging_freq': 7}. Best is trial 27 with value: 0.6882582742878182.
bagging, val_score: 0.688258: 60%|###### | 6/10 [00:06<00:04, 1.15s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008379 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685889 valid's binary_logloss: 0.68974
[200] train's binary_logloss: 0.682644 valid's binary_logloss: 0.68847
[300] train's binary_logloss: 0.679926 valid's binary_logloss: 0.688667
Early stopping, best iteration is:
[200] train's binary_logloss: 0.682644 valid's binary_logloss: 0.68847
bagging, val_score: 0.688258: 70%|####### | 7/10 [00:08<00:03, 1.14s/it][I 2020-09-27 05:02:37,721] Trial 33 finished with value: 0.6884699706432157 and parameters: {'bagging_fraction': 0.6161422919690354, 'bagging_freq': 7}. Best is trial 27 with value: 0.6882582742878182.
bagging, val_score: 0.688258: 70%|####### | 7/10 [00:08<00:03, 1.14s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013266 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.686069 valid's binary_logloss: 0.689747
[200] train's binary_logloss: 0.682773 valid's binary_logloss: 0.688633
[300] train's binary_logloss: 0.679937 valid's binary_logloss: 0.68895
Early stopping, best iteration is:
[205] train's binary_logloss: 0.682645 valid's binary_logloss: 0.688588
bagging, val_score: 0.688258: 80%|######## | 8/10 [00:09<00:02, 1.14s/it][I 2020-09-27 05:02:38,853] Trial 34 finished with value: 0.6885878332172652 and parameters: {'bagging_fraction': 0.6004490251930547, 'bagging_freq': 7}. Best is trial 27 with value: 0.6882582742878182.
bagging, val_score: 0.688258: 80%|######## | 8/10 [00:09<00:02, 1.14s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008258 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.686065 valid's binary_logloss: 0.689456
[200] train's binary_logloss: 0.682802 valid's binary_logloss: 0.688959
[300] train's binary_logloss: 0.680026 valid's binary_logloss: 0.688847
Early stopping, best iteration is:
[262] train's binary_logloss: 0.681101 valid's binary_logloss: 0.688619
bagging, val_score: 0.688258: 90%|######### | 9/10 [00:10<00:01, 1.21s/it][I 2020-09-27 05:02:40,228] Trial 35 finished with value: 0.6886193436195897 and parameters: {'bagging_fraction': 0.7777960275374113, 'bagging_freq': 7}. Best is trial 27 with value: 0.6882582742878182.
bagging, val_score: 0.688258: 90%|######### | 9/10 [00:10<00:01, 1.21s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008496 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685807 valid's binary_logloss: 0.689065
[200] train's binary_logloss: 0.68264 valid's binary_logloss: 0.6887
[300] train's binary_logloss: 0.680006 valid's binary_logloss: 0.688483
Early stopping, best iteration is:
[291] train's binary_logloss: 0.680267 valid's binary_logloss: 0.688364
bagging, val_score: 0.688258: 100%|##########| 10/10 [00:11<00:00, 1.25s/it][I 2020-09-27 05:02:41,562] Trial 36 finished with value: 0.6883635702761751 and parameters: {'bagging_fraction': 0.4580982361208543, 'bagging_freq': 4}. Best is trial 27 with value: 0.6882582742878182.
bagging, val_score: 0.688258: 100%|##########| 10/10 [00:11<00:00, 1.19s/it]
feature_fraction_stage2, val_score: 0.688258: 0%| | 0/6 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008209 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486
[200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315
[300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232
[400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467
Early stopping, best iteration is:
[334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072
feature_fraction_stage2, val_score: 0.688072: 17%|#6 | 1/6 [00:01<00:07, 1.55s/it][I 2020-09-27 05:02:43,127] Trial 37 finished with value: 0.6880722259887746 and parameters: {'feature_fraction': 0.748}. Best is trial 37 with value: 0.6880722259887746.
feature_fraction_stage2, val_score: 0.688072: 17%|#6 | 1/6 [00:01<00:07, 1.55s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.003567 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486
[200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315
[300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232
[400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467
Early stopping, best iteration is:
[334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072
feature_fraction_stage2, val_score: 0.688072: 33%|###3 | 2/6 [00:03<00:06, 1.58s/it][I 2020-09-27 05:02:44,774] Trial 38 finished with value: 0.6880722259887746 and parameters: {'feature_fraction': 0.716}. Best is trial 37 with value: 0.6880722259887746.
feature_fraction_stage2, val_score: 0.688072: 33%|###3 | 2/6 [00:03<00:06, 1.58s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008780 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.686006 valid's binary_logloss: 0.689628
[200] train's binary_logloss: 0.682779 valid's binary_logloss: 0.688355
[300] train's binary_logloss: 0.68005 valid's binary_logloss: 0.688419
Early stopping, best iteration is:
[219] train's binary_logloss: 0.682239 valid's binary_logloss: 0.688258
feature_fraction_stage2, val_score: 0.688072: 50%|##### | 3/6 [00:04<00:04, 1.45s/it][I 2020-09-27 05:02:45,923] Trial 39 finished with value: 0.6882582742878182 and parameters: {'feature_fraction': 0.6839999999999999}. Best is trial 37 with value: 0.6880722259887746.
feature_fraction_stage2, val_score: 0.688072: 50%|##### | 3/6 [00:04<00:04, 1.45s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009262 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685951 valid's binary_logloss: 0.689912
[200] train's binary_logloss: 0.682613 valid's binary_logloss: 0.688642
[300] train's binary_logloss: 0.679824 valid's binary_logloss: 0.688952
Early stopping, best iteration is:
[201] train's binary_logloss: 0.682582 valid's binary_logloss: 0.688618
feature_fraction_stage2, val_score: 0.688072: 67%|######6 | 4/6 [00:05<00:02, 1.36s/it][I 2020-09-27 05:02:47,065] Trial 40 finished with value: 0.6886180547556369 and parameters: {'feature_fraction': 0.7799999999999999}. Best is trial 37 with value: 0.6880722259887746.
feature_fraction_stage2, val_score: 0.688072: 67%|######6 | 4/6 [00:05<00:02, 1.36s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009195 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685997 valid's binary_logloss: 0.689645
[200] train's binary_logloss: 0.682772 valid's binary_logloss: 0.688789
[300] train's binary_logloss: 0.680042 valid's binary_logloss: 0.689271
Early stopping, best iteration is:
[200] train's binary_logloss: 0.682772 valid's binary_logloss: 0.688789
feature_fraction_stage2, val_score: 0.688072: 83%|########3 | 5/6 [00:06<00:01, 1.29s/it][I 2020-09-27 05:02:48,184] Trial 41 finished with value: 0.6887892049667754 and parameters: {'feature_fraction': 0.62}. Best is trial 37 with value: 0.6880722259887746.
feature_fraction_stage2, val_score: 0.688072: 83%|########3 | 5/6 [00:06<00:01, 1.29s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013986 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.686063 valid's binary_logloss: 0.68975
[200] train's binary_logloss: 0.682823 valid's binary_logloss: 0.688614
[300] train's binary_logloss: 0.680059 valid's binary_logloss: 0.689073
Early stopping, best iteration is:
[206] train's binary_logloss: 0.68266 valid's binary_logloss: 0.688579
feature_fraction_stage2, val_score: 0.688072: 100%|##########| 6/6 [00:07<00:00, 1.24s/it][I 2020-09-27 05:02:49,309] Trial 42 finished with value: 0.688578931631993 and parameters: {'feature_fraction': 0.652}. Best is trial 37 with value: 0.6880722259887746.
feature_fraction_stage2, val_score: 0.688072: 100%|##########| 6/6 [00:07<00:00, 1.29s/it]
regularization_factors, val_score: 0.688072: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001438 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.686053 valid's binary_logloss: 0.689655
[200] train's binary_logloss: 0.682817 valid's binary_logloss: 0.688251
[300] train's binary_logloss: 0.680044 valid's binary_logloss: 0.688312
Early stopping, best iteration is:
[274] train's binary_logloss: 0.68083 valid's binary_logloss: 0.68815
regularization_factors, val_score: 0.688072: 5%|5 | 1/20 [00:01<00:26, 1.40s/it][I 2020-09-27 05:02:50,722] Trial 43 finished with value: 0.6881495198006506 and parameters: {'lambda_l1': 0.0485944386213443, 'lambda_l2': 0.24935601021074602}. Best is trial 43 with value: 0.6881495198006506.
regularization_factors, val_score: 0.688072: 5%|5 | 1/20 [00:01<00:26, 1.40s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001480 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68603 valid's binary_logloss: 0.689695
[200] train's binary_logloss: 0.682787 valid's binary_logloss: 0.688424
[300] train's binary_logloss: 0.680205 valid's binary_logloss: 0.688656
Early stopping, best iteration is:
[221] train's binary_logloss: 0.682216 valid's binary_logloss: 0.688392
regularization_factors, val_score: 0.688072: 10%|# | 2/20 [00:02<00:24, 1.35s/it][I 2020-09-27 05:02:51,960] Trial 44 finished with value: 0.6883919982325207 and parameters: {'lambda_l1': 0.10751881913962534, 'lambda_l2': 0.5824729388634683}. Best is trial 43 with value: 0.6881495198006506.
regularization_factors, val_score: 0.688072: 10%|# | 2/20 [00:02<00:24, 1.35s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008830 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689487
[200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315
[300] train's binary_logloss: 0.679962 valid's binary_logloss: 0.688232
[400] train's binary_logloss: 0.677326 valid's binary_logloss: 0.688466
Early stopping, best iteration is:
[334] train's binary_logloss: 0.679068 valid's binary_logloss: 0.688072
regularization_factors, val_score: 0.688072: 15%|#5 | 3/20 [00:04<00:24, 1.41s/it][I 2020-09-27 05:02:53,524] Trial 45 finished with value: 0.6880722890783393 and parameters: {'lambda_l1': 1.188253884167047e-06, 'lambda_l2': 0.005588276816712126}. Best is trial 45 with value: 0.6880722890783393.
regularization_factors, val_score: 0.688072: 15%|#5 | 3/20 [00:04<00:24, 1.41s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002086 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689487
[200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315
[300] train's binary_logloss: 0.679962 valid's binary_logloss: 0.688232
[400] train's binary_logloss: 0.677326 valid's binary_logloss: 0.688466
Early stopping, best iteration is:
[334] train's binary_logloss: 0.679068 valid's binary_logloss: 0.688072
regularization_factors, val_score: 0.688072: 20%|## | 4/20 [00:05<00:23, 1.48s/it][I 2020-09-27 05:02:55,155] Trial 46 finished with value: 0.6880722851297056 and parameters: {'lambda_l1': 5.105795902683296e-08, 'lambda_l2': 0.005237582202723369}. Best is trial 46 with value: 0.6880722851297056.
regularization_factors, val_score: 0.688072: 20%|## | 4/20 [00:05<00:23, 1.48s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001606 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689487
[200] train's binary_logloss: 0.682685 valid's binary_logloss: 0.688315
[300] train's binary_logloss: 0.679962 valid's binary_logloss: 0.688232
[400] train's binary_logloss: 0.67738 valid's binary_logloss: 0.68864
Early stopping, best iteration is:
[330] train's binary_logloss: 0.679165 valid's binary_logloss: 0.688085
regularization_factors, val_score: 0.688072: 25%|##5 | 5/20 [00:07<00:22, 1.52s/it][I 2020-09-27 05:02:56,785] Trial 47 finished with value: 0.6880848414842485 and parameters: {'lambda_l1': 1.720232956611629e-08, 'lambda_l2': 0.007516638600538305}. Best is trial 46 with value: 0.6880722851297056.
regularization_factors, val_score: 0.688072: 25%|##5 | 5/20 [00:07<00:22, 1.52s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.014048 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486
[200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315
[300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232
[400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467
Early stopping, best iteration is:
[334] train's binary_logloss: 0.679067 valid's binary_logloss: 0.688072
regularization_factors, val_score: 0.688072: 30%|### | 6/20 [00:09<00:21, 1.54s/it][I 2020-09-27 05:02:58,355] Trial 48 finished with value: 0.688072230112268 and parameters: {'lambda_l1': 1.027357618260445e-08, 'lambda_l2': 0.00036547314253742376}. Best is trial 48 with value: 0.688072230112268.
regularization_factors, val_score: 0.688072: 30%|### | 6/20 [00:09<00:21, 1.54s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008658 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486
[200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315
[300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232
[400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467
Early stopping, best iteration is:
[334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072
regularization_factors, val_score: 0.688072: 35%|###5 | 7/20 [00:10<00:20, 1.55s/it][I 2020-09-27 05:02:59,927] Trial 49 finished with value: 0.6880722260400305 and parameters: {'lambda_l1': 2.5990203700513686e-08, 'lambda_l2': 4.6609028631019925e-06}. Best is trial 49 with value: 0.6880722260400305.
regularization_factors, val_score: 0.688072: 35%|###5 | 7/20 [00:10<00:20, 1.55s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001464 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486
[200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315
[300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232
[400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467
Early stopping, best iteration is:
[334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072
regularization_factors, val_score: 0.688072: 40%|#### | 8/20 [00:12<00:18, 1.56s/it][I 2020-09-27 05:03:01,518] Trial 50 finished with value: 0.6880722259938851 and parameters: {'lambda_l1': 1.0501739122696127e-08, 'lambda_l2': 2.906436684517273e-07}. Best is trial 50 with value: 0.6880722259938851.
regularization_factors, val_score: 0.688072: 40%|#### | 8/20 [00:12<00:18, 1.56s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008492 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486
[200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315
[300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232
[400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467
Early stopping, best iteration is:
[334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072
regularization_factors, val_score: 0.688072: 45%|####5 | 9/20 [00:13<00:17, 1.56s/it][I 2020-09-27 05:03:03,084] Trial 51 finished with value: 0.6880722259917854 and parameters: {'lambda_l1': 1.0110083939507621e-08, 'lambda_l2': 1.8869898459706898e-07}. Best is trial 51 with value: 0.6880722259917854.
regularization_factors, val_score: 0.688072: 45%|####5 | 9/20 [00:13<00:17, 1.56s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008606 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486
[200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315
[300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232
[400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467
Early stopping, best iteration is:
[334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072
regularization_factors, val_score: 0.688072: 50%|##### | 10/20 [00:15<00:15, 1.56s/it][I 2020-09-27 05:03:04,652] Trial 52 finished with value: 0.6880722259908346 and parameters: {'lambda_l1': 1.5445331505275187e-08, 'lambda_l2': 1.2613210373316155e-07}. Best is trial 52 with value: 0.6880722259908346.
regularization_factors, val_score: 0.688072: 50%|##### | 10/20 [00:15<00:15, 1.56s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008206 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486
[200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315
[300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232
[400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467
Early stopping, best iteration is:
[334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072
regularization_factors, val_score: 0.688072: 55%|#####5 | 11/20 [00:16<00:14, 1.56s/it][I 2020-09-27 05:03:06,216] Trial 53 finished with value: 0.6880722259908322 and parameters: {'lambda_l1': 1.187725756223345e-08, 'lambda_l2': 8.260605690107965e-08}. Best is trial 53 with value: 0.6880722259908322.
regularization_factors, val_score: 0.688072: 55%|#####5 | 11/20 [00:16<00:14, 1.56s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001731 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486
[200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315
[300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232
[400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467
Early stopping, best iteration is:
[334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072
regularization_factors, val_score: 0.688072: 60%|###### | 12/20 [00:18<00:12, 1.59s/it][I 2020-09-27 05:03:07,867] Trial 54 finished with value: 0.6880722259624853 and parameters: {'lambda_l1': 2.1107731769847976e-06, 'lambda_l2': 1.024124993705147e-08}. Best is trial 54 with value: 0.6880722259624853.
regularization_factors, val_score: 0.688072: 60%|###### | 12/20 [00:18<00:12, 1.59s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013514 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486
[200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315
[300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232
[400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467
Early stopping, best iteration is:
[334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072
regularization_factors, val_score: 0.688072: 65%|######5 | 13/20 [00:20<00:11, 1.59s/it][I 2020-09-27 05:03:09,440] Trial 55 finished with value: 0.6880722259228028 and parameters: {'lambda_l1': 4.841910937380013e-06, 'lambda_l2': 1.6262995524915506e-08}. Best is trial 55 with value: 0.6880722259228028.
regularization_factors, val_score: 0.688072: 65%|######5 | 13/20 [00:20<00:11, 1.59s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002085 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486
[200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315
[300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232
[400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467
Early stopping, best iteration is:
[334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072
regularization_factors, val_score: 0.688072: 70%|####### | 14/20 [00:21<00:09, 1.60s/it][I 2020-09-27 05:03:11,073] Trial 56 finished with value: 0.6880722257360149 and parameters: {'lambda_l1': 1.964871251332563e-05, 'lambda_l2': 1.3890624113900224e-08}. Best is trial 56 with value: 0.6880722257360149.
regularization_factors, val_score: 0.688072: 70%|####### | 14/20 [00:21<00:09, 1.60s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013271 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486
[200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315
[300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232
[400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467
Early stopping, best iteration is:
[334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072
regularization_factors, val_score: 0.688072: 75%|#######5 | 15/20 [00:23<00:07, 1.59s/it][I 2020-09-27 05:03:12,639] Trial 57 finished with value: 0.6880722255253509 and parameters: {'lambda_l1': 3.358919485655909e-05, 'lambda_l2': 1.6266181828610483e-08}. Best is trial 57 with value: 0.6880722255253509.
regularization_factors, val_score: 0.688072: 75%|#######5 | 15/20 [00:23<00:07, 1.59s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009828 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486
[200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315
[300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232
[400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467
Early stopping, best iteration is:
[334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072
regularization_factors, val_score: 0.688072: 80%|######## | 16/20 [00:24<00:06, 1.58s/it][I 2020-09-27 05:03:14,206] Trial 58 finished with value: 0.6880722255987741 and parameters: {'lambda_l1': 2.8108351646575752e-05, 'lambda_l2': 1.100109030564083e-08}. Best is trial 57 with value: 0.6880722255253509.
regularization_factors, val_score: 0.688072: 80%|######## | 16/20 [00:24<00:06, 1.58s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012705 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486
[200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315
[300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232
[400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467
Early stopping, best iteration is:
[334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072
regularization_factors, val_score: 0.688072: 85%|########5 | 17/20 [00:26<00:04, 1.57s/it][I 2020-09-27 05:03:15,742] Trial 59 finished with value: 0.6880722255024915 and parameters: {'lambda_l1': 3.5103404293664e-05, 'lambda_l2': 1.005446328322094e-08}. Best is trial 59 with value: 0.6880722255024915.
regularization_factors, val_score: 0.688072: 85%|########5 | 17/20 [00:26<00:04, 1.57s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001474 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486
[200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315
[300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232
[400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467
Early stopping, best iteration is:
[334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072
regularization_factors, val_score: 0.688072: 90%|######### | 18/20 [00:28<00:03, 1.58s/it][I 2020-09-27 05:03:17,361] Trial 60 finished with value: 0.6880722254979789 and parameters: {'lambda_l1': 3.541626511270535e-05, 'lambda_l2': 1.0717975785657199e-08}. Best is trial 60 with value: 0.6880722254979789.
regularization_factors, val_score: 0.688072: 90%|######### | 18/20 [00:28<00:03, 1.58s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010965 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486
[200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315
[300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232
[400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467
Early stopping, best iteration is:
[334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072
regularization_factors, val_score: 0.688072: 95%|#########5| 19/20 [00:29<00:01, 1.58s/it][I 2020-09-27 05:03:18,921] Trial 61 finished with value: 0.6880722256490812 and parameters: {'lambda_l1': 2.4664030202905467e-05, 'lambda_l2': 1.0803279789819484e-08}. Best is trial 60 with value: 0.6880722254979789.
regularization_factors, val_score: 0.688072: 95%|#########5| 19/20 [00:29<00:01, 1.58s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008152 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486
[200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315
[300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232
[400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467
Early stopping, best iteration is:
[334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072
regularization_factors, val_score: 0.688072: 100%|##########| 20/20 [00:31<00:00, 1.56s/it][I 2020-09-27 05:03:20,457] Trial 62 finished with value: 0.6880722251720996 and parameters: {'lambda_l1': 5.8781935567800015e-05, 'lambda_l2': 1.1695290416442605e-08}. Best is trial 62 with value: 0.6880722251720996.
regularization_factors, val_score: 0.688072: 100%|##########| 20/20 [00:31<00:00, 1.56s/it]
min_data_in_leaf, val_score: 0.688072: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012689 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.686108 valid's binary_logloss: 0.689957
[200] train's binary_logloss: 0.683126 valid's binary_logloss: 0.688938
Early stopping, best iteration is:
[199] train's binary_logloss: 0.683149 valid's binary_logloss: 0.688886
min_data_in_leaf, val_score: 0.688072: 20%|## | 1/5 [00:01<00:04, 1.14s/it][I 2020-09-27 05:03:21,608] Trial 63 finished with value: 0.6888860610141573 and parameters: {'min_child_samples': 100}. Best is trial 63 with value: 0.6888860610141573.
min_data_in_leaf, val_score: 0.688072: 20%|## | 1/5 [00:01<00:04, 1.14s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009121 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685963 valid's binary_logloss: 0.689803
[200] train's binary_logloss: 0.682736 valid's binary_logloss: 0.688538
Early stopping, best iteration is:
[198] train's binary_logloss: 0.682793 valid's binary_logloss: 0.688516
min_data_in_leaf, val_score: 0.688072: 40%|#### | 2/5 [00:02<00:03, 1.14s/it][I 2020-09-27 05:03:22,746] Trial 64 finished with value: 0.6885159620293074 and parameters: {'min_child_samples': 25}. Best is trial 64 with value: 0.6885159620293074.
min_data_in_leaf, val_score: 0.688072: 40%|#### | 2/5 [00:02<00:03, 1.14s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008620 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68607 valid's binary_logloss: 0.689662
[200] train's binary_logloss: 0.682938 valid's binary_logloss: 0.688496
[300] train's binary_logloss: 0.68032 valid's binary_logloss: 0.688457
[400] train's binary_logloss: 0.677831 valid's binary_logloss: 0.688297
Early stopping, best iteration is:
[330] train's binary_logloss: 0.679568 valid's binary_logloss: 0.6881
min_data_in_leaf, val_score: 0.688072: 60%|###### | 3/5 [00:03<00:02, 1.27s/it][I 2020-09-27 05:03:24,343] Trial 65 finished with value: 0.6880998602504527 and parameters: {'min_child_samples': 50}. Best is trial 65 with value: 0.6880998602504527.
min_data_in_leaf, val_score: 0.688072: 60%|###### | 3/5 [00:03<00:02, 1.27s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001443 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685989 valid's binary_logloss: 0.689573
[200] train's binary_logloss: 0.682702 valid's binary_logloss: 0.688452
[300] train's binary_logloss: 0.679934 valid's binary_logloss: 0.68856
Early stopping, best iteration is:
[272] train's binary_logloss: 0.680746 valid's binary_logloss: 0.688417
min_data_in_leaf, val_score: 0.688072: 80%|######## | 4/5 [00:05<00:01, 1.32s/it][I 2020-09-27 05:03:25,758] Trial 66 finished with value: 0.6884168700112442 and parameters: {'min_child_samples': 5}. Best is trial 65 with value: 0.6880998602504527.
min_data_in_leaf, val_score: 0.688072: 80%|######## | 4/5 [00:05<00:01, 1.32s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008044 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.686007 valid's binary_logloss: 0.689624
[200] train's binary_logloss: 0.682678 valid's binary_logloss: 0.688328
[300] train's binary_logloss: 0.679893 valid's binary_logloss: 0.688247
[400] train's binary_logloss: 0.677249 valid's binary_logloss: 0.688678
Early stopping, best iteration is:
[311] train's binary_logloss: 0.679598 valid's binary_logloss: 0.688186
min_data_in_leaf, val_score: 0.688072: 100%|##########| 5/5 [00:06<00:00, 1.38s/it][I 2020-09-27 05:03:27,278] Trial 67 finished with value: 0.6881857897062641 and parameters: {'min_child_samples': 10}. Best is trial 65 with value: 0.6880998602504527.
min_data_in_leaf, val_score: 0.688072: 100%|##########| 5/5 [00:06<00:00, 1.36s/it]
Fold : 9
[I 2020-09-27 05:03:27,416] A new study created in memory with name: no-name-e6a1e2b2-b26c-4a35-b26d-98523bba8f2b
feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001148 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664968 valid's binary_logloss: 0.689984
Early stopping, best iteration is:
[38] train's binary_logloss: 0.679101 valid's binary_logloss: 0.689149
feature_fraction, val_score: 0.689149: 14%|#4 | 1/7 [00:01<00:06, 1.12s/it][I 2020-09-27 05:03:28,547] Trial 0 finished with value: 0.6891486028436887 and parameters: {'feature_fraction': 0.5}. Best is trial 0 with value: 0.6891486028436887.
feature_fraction, val_score: 0.689149: 14%|#4 | 1/7 [00:01<00:06, 1.12s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.014784 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.662784 valid's binary_logloss: 0.690663
Early stopping, best iteration is:
[40] train's binary_logloss: 0.677488 valid's binary_logloss: 0.689489
feature_fraction, val_score: 0.689149: 29%|##8 | 2/7 [00:02<00:05, 1.19s/it][I 2020-09-27 05:03:29,885] Trial 1 finished with value: 0.6894894237035651 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 0 with value: 0.6891486028436887.
feature_fraction, val_score: 0.689149: 29%|##8 | 2/7 [00:02<00:05, 1.19s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000849 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.66599 valid's binary_logloss: 0.69036
Early stopping, best iteration is:
[64] train's binary_logloss: 0.67328 valid's binary_logloss: 0.689617
feature_fraction, val_score: 0.689149: 43%|####2 | 3/7 [00:03<00:04, 1.09s/it][I 2020-09-27 05:03:30,757] Trial 2 finished with value: 0.6896170710885305 and parameters: {'feature_fraction': 0.4}. Best is trial 0 with value: 0.6891486028436887.
feature_fraction, val_score: 0.689149: 43%|####2 | 3/7 [00:03<00:04, 1.09s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.005632 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.66269 valid's binary_logloss: 0.690688
Early stopping, best iteration is:
[38] train's binary_logloss: 0.67783 valid's binary_logloss: 0.689711
feature_fraction, val_score: 0.689149: 57%|#####7 | 4/7 [00:04<00:03, 1.03s/it][I 2020-09-27 05:03:31,639] Trial 3 finished with value: 0.6897107894464999 and parameters: {'feature_fraction': 1.0}. Best is trial 0 with value: 0.6891486028436887.
feature_fraction, val_score: 0.689149: 57%|#####7 | 4/7 [00:04<00:03, 1.03s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013949 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.662985 valid's binary_logloss: 0.690369
Early stopping, best iteration is:
[89] train's binary_logloss: 0.665582 valid's binary_logloss: 0.689673
feature_fraction, val_score: 0.689149: 71%|#######1 | 5/7 [00:05<00:02, 1.02s/it][I 2020-09-27 05:03:32,641] Trial 4 finished with value: 0.6896732000011675 and parameters: {'feature_fraction': 0.8}. Best is trial 0 with value: 0.6891486028436887.
feature_fraction, val_score: 0.689149: 71%|#######1 | 5/7 [00:05<00:02, 1.02s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007569 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.663886 valid's binary_logloss: 0.690174
Early stopping, best iteration is:
[61] train's binary_logloss: 0.672601 valid's binary_logloss: 0.689421
feature_fraction, val_score: 0.689149: 86%|########5 | 6/7 [00:06<00:00, 1.04it/s][I 2020-09-27 05:03:33,477] Trial 5 finished with value: 0.6894210990140921 and parameters: {'feature_fraction': 0.7}. Best is trial 0 with value: 0.6891486028436887.
feature_fraction, val_score: 0.689149: 86%|########5 | 6/7 [00:06<00:00, 1.04it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008826 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.664317 valid's binary_logloss: 0.690145
Early stopping, best iteration is:
[52] train's binary_logloss: 0.675207 valid's binary_logloss: 0.689455
feature_fraction, val_score: 0.689149: 100%|##########| 7/7 [00:06<00:00, 1.11it/s][I 2020-09-27 05:03:34,239] Trial 6 finished with value: 0.6894553081219641 and parameters: {'feature_fraction': 0.6}. Best is trial 0 with value: 0.6891486028436887.
feature_fraction, val_score: 0.689149: 100%|##########| 7/7 [00:06<00:00, 1.03it/s]
num_leaves, val_score: 0.689149: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000950 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.564202 valid's binary_logloss: 0.694875
Early stopping, best iteration is:
[19] train's binary_logloss: 0.658179 valid's binary_logloss: 0.69083
num_leaves, val_score: 0.689149: 5%|5 | 1/20 [00:01<00:23, 1.25s/it][I 2020-09-27 05:03:35,508] Trial 7 finished with value: 0.6908299824401397 and parameters: {'num_leaves': 204}. Best is trial 7 with value: 0.6908299824401397.
num_leaves, val_score: 0.689149: 5%|5 | 1/20 [00:01<00:23, 1.25s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000951 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.564078 valid's binary_logloss: 0.696783
Early stopping, best iteration is:
[13] train's binary_logloss: 0.668565 valid's binary_logloss: 0.690921
num_leaves, val_score: 0.689149: 10%|# | 2/20 [00:02<00:22, 1.23s/it][I 2020-09-27 05:03:36,684] Trial 8 finished with value: 0.6909207765265465 and parameters: {'num_leaves': 202}. Best is trial 7 with value: 0.6908299824401397.
num_leaves, val_score: 0.689149: 10%|# | 2/20 [00:02<00:22, 1.23s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000865 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.625795 valid's binary_logloss: 0.693319
Early stopping, best iteration is:
[35] train's binary_logloss: 0.663313 valid's binary_logloss: 0.690147
num_leaves, val_score: 0.689149: 15%|#5 | 3/20 [00:03<00:19, 1.13s/it][I 2020-09-27 05:03:37,591] Trial 9 finished with value: 0.6901469432883655 and parameters: {'num_leaves': 89}. Best is trial 9 with value: 0.6901469432883655.
num_leaves, val_score: 0.689149: 15%|#5 | 3/20 [00:03<00:19, 1.13s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000928 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.690045 valid's binary_logloss: 0.690754
[200] train's binary_logloss: 0.688849 valid's binary_logloss: 0.690081
[300] train's binary_logloss: 0.688099 valid's binary_logloss: 0.6897
[400] train's binary_logloss: 0.687569 valid's binary_logloss: 0.689504
[500] train's binary_logloss: 0.687173 valid's binary_logloss: 0.689388
[600] train's binary_logloss: 0.686865 valid's binary_logloss: 0.689372
Early stopping, best iteration is:
[537] train's binary_logloss: 0.687051 valid's binary_logloss: 0.689339
num_leaves, val_score: 0.689149: 20%|## | 4/20 [00:05<00:22, 1.41s/it][I 2020-09-27 05:03:39,636] Trial 10 finished with value: 0.6893392852593333 and parameters: {'num_leaves': 2}. Best is trial 10 with value: 0.6893392852593333.
num_leaves, val_score: 0.689149: 20%|## | 4/20 [00:05<00:22, 1.41s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000928 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.686226 valid's binary_logloss: 0.689731
[200] train's binary_logloss: 0.683165 valid's binary_logloss: 0.689426
[300] train's binary_logloss: 0.680433 valid's binary_logloss: 0.689593
Early stopping, best iteration is:
[221] train's binary_logloss: 0.682545 valid's binary_logloss: 0.689351
num_leaves, val_score: 0.689149: 25%|##5 | 5/20 [00:06<00:20, 1.34s/it][I 2020-09-27 05:03:40,838] Trial 11 finished with value: 0.6893514061615624 and parameters: {'num_leaves': 5}. Best is trial 10 with value: 0.6893392852593333.
num_leaves, val_score: 0.689149: 25%|##5 | 5/20 [00:06<00:20, 1.34s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001034 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.677343 valid's binary_logloss: 0.689733
Early stopping, best iteration is:
[73] train's binary_logloss: 0.680397 valid's binary_logloss: 0.689164
num_leaves, val_score: 0.689149: 30%|### | 6/20 [00:07<00:16, 1.18s/it][I 2020-09-27 05:03:41,643] Trial 12 finished with value: 0.6891638627451019 and parameters: {'num_leaves': 15}. Best is trial 12 with value: 0.6891638627451019.
num_leaves, val_score: 0.689149: 30%|### | 6/20 [00:07<00:16, 1.18s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000978 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.625795 valid's binary_logloss: 0.693319
Early stopping, best iteration is:
[35] train's binary_logloss: 0.663313 valid's binary_logloss: 0.690147
num_leaves, val_score: 0.689149: 35%|###5 | 7/20 [00:08<00:14, 1.12s/it][I 2020-09-27 05:03:42,610] Trial 13 finished with value: 0.6901469432883656 and parameters: {'num_leaves': 89}. Best is trial 12 with value: 0.6891638627451019.
num_leaves, val_score: 0.689149: 35%|###5 | 7/20 [00:08<00:14, 1.12s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000900 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.645429 valid's binary_logloss: 0.691128
Early stopping, best iteration is:
[37] train's binary_logloss: 0.670937 valid's binary_logloss: 0.689723
num_leaves, val_score: 0.689149: 40%|#### | 8/20 [00:09<00:12, 1.03s/it][I 2020-09-27 05:03:43,427] Trial 14 finished with value: 0.6897233852217974 and parameters: {'num_leaves': 58}. Best is trial 12 with value: 0.6891638627451019.
num_leaves, val_score: 0.689149: 40%|#### | 8/20 [00:09<00:12, 1.03s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000972 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.53862 valid's binary_logloss: 0.696319
Early stopping, best iteration is:
[19] train's binary_logloss: 0.651313 valid's binary_logloss: 0.690486
num_leaves, val_score: 0.689149: 45%|####5 | 9/20 [00:10<00:12, 1.13s/it][I 2020-09-27 05:03:44,806] Trial 15 finished with value: 0.6904857400051433 and parameters: {'num_leaves': 254}. Best is trial 12 with value: 0.6891638627451019.
num_leaves, val_score: 0.689149: 45%|####5 | 9/20 [00:10<00:12, 1.13s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000921 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.58509 valid's binary_logloss: 0.697538
Early stopping, best iteration is:
[18] train's binary_logloss: 0.665762 valid's binary_logloss: 0.690884
num_leaves, val_score: 0.689149: 50%|##### | 10/20 [00:11<00:11, 1.10s/it][I 2020-09-27 05:03:45,833] Trial 16 finished with value: 0.6908843676817171 and parameters: {'num_leaves': 160}. Best is trial 12 with value: 0.6891638627451019.
num_leaves, val_score: 0.689149: 50%|##### | 10/20 [00:11<00:11, 1.10s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008682 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.655506 valid's binary_logloss: 0.690438
Early stopping, best iteration is:
[34] train's binary_logloss: 0.67629 valid's binary_logloss: 0.689323
num_leaves, val_score: 0.689149: 55%|#####5 | 11/20 [00:12<00:08, 1.03it/s][I 2020-09-27 05:03:46,509] Trial 17 finished with value: 0.6893229030060266 and parameters: {'num_leaves': 44}. Best is trial 12 with value: 0.6891638627451019.
num_leaves, val_score: 0.689149: 55%|#####5 | 11/20 [00:12<00:08, 1.03it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000877 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.598808 valid's binary_logloss: 0.694092
Early stopping, best iteration is:
[16] train's binary_logloss: 0.671679 valid's binary_logloss: 0.690451
num_leaves, val_score: 0.689149: 60%|###### | 12/20 [00:13<00:07, 1.02it/s][I 2020-09-27 05:03:47,497] Trial 18 finished with value: 0.6904513066474659 and parameters: {'num_leaves': 136}. Best is trial 12 with value: 0.6891638627451019.
num_leaves, val_score: 0.689149: 60%|###### | 12/20 [00:13<00:07, 1.02it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000956 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.665649 valid's binary_logloss: 0.690234
Early stopping, best iteration is:
[38] train's binary_logloss: 0.679544 valid's binary_logloss: 0.689555
num_leaves, val_score: 0.689149: 65%|######5 | 13/20 [00:13<00:06, 1.12it/s][I 2020-09-27 05:03:48,195] Trial 19 finished with value: 0.6895546123306635 and parameters: {'num_leaves': 30}. Best is trial 12 with value: 0.6891638627451019.
num_leaves, val_score: 0.689149: 65%|######5 | 13/20 [00:13<00:06, 1.12it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000933 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.625105 valid's binary_logloss: 0.692852
Early stopping, best iteration is:
[37] train's binary_logloss: 0.661371 valid's binary_logloss: 0.690145
num_leaves, val_score: 0.689149: 70%|####### | 14/20 [00:14<00:05, 1.09it/s][I 2020-09-27 05:03:49,159] Trial 20 finished with value: 0.6901453602236771 and parameters: {'num_leaves': 91}. Best is trial 12 with value: 0.6891638627451019.
num_leaves, val_score: 0.689149: 70%|####### | 14/20 [00:14<00:05, 1.09it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000954 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.658357 valid's binary_logloss: 0.69009
Early stopping, best iteration is:
[53] train's binary_logloss: 0.671442 valid's binary_logloss: 0.689611
num_leaves, val_score: 0.689149: 75%|#######5 | 15/20 [00:15<00:04, 1.12it/s][I 2020-09-27 05:03:49,991] Trial 21 finished with value: 0.6896105933356692 and parameters: {'num_leaves': 40}. Best is trial 12 with value: 0.6891638627451019.
num_leaves, val_score: 0.689149: 75%|#######5 | 15/20 [00:15<00:04, 1.12it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000936 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685253 valid's binary_logloss: 0.689642
[200] train's binary_logloss: 0.681379 valid's binary_logloss: 0.689235
Early stopping, best iteration is:
[185] train's binary_logloss: 0.681921 valid's binary_logloss: 0.689088
num_leaves, val_score: 0.689088: 80%|######## | 16/20 [00:16<00:03, 1.04it/s][I 2020-09-27 05:03:51,133] Trial 22 finished with value: 0.689088142033773 and parameters: {'num_leaves': 6}. Best is trial 22 with value: 0.689088142033773.
num_leaves, val_score: 0.689088: 80%|######## | 16/20 [00:16<00:03, 1.04it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000953 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.687295 valid's binary_logloss: 0.68994
[200] train's binary_logloss: 0.684807 valid's binary_logloss: 0.689595
[300] train's binary_logloss: 0.682867 valid's binary_logloss: 0.689608
Early stopping, best iteration is:
[256] train's binary_logloss: 0.683708 valid's binary_logloss: 0.689506
num_leaves, val_score: 0.689088: 85%|########5 | 17/20 [00:18<00:03, 1.06s/it][I 2020-09-27 05:03:52,425] Trial 23 finished with value: 0.6895057052158261 and parameters: {'num_leaves': 4}. Best is trial 22 with value: 0.689088142033773.
num_leaves, val_score: 0.689088: 85%|########5 | 17/20 [00:18<00:03, 1.06s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011377 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.636575 valid's binary_logloss: 0.690665
Early stopping, best iteration is:
[36] train's binary_logloss: 0.667345 valid's binary_logloss: 0.689716
num_leaves, val_score: 0.689088: 90%|######### | 18/20 [00:18<00:01, 1.01it/s][I 2020-09-27 05:03:53,255] Trial 24 finished with value: 0.6897160545946952 and parameters: {'num_leaves': 72}. Best is trial 22 with value: 0.689088142033773.
num_leaves, val_score: 0.689088: 90%|######### | 18/20 [00:19<00:01, 1.01it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000933 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.667234 valid's binary_logloss: 0.689986
Early stopping, best iteration is:
[74] train's binary_logloss: 0.672163 valid's binary_logloss: 0.689426
num_leaves, val_score: 0.689088: 95%|#########5| 19/20 [00:19<00:00, 1.05it/s][I 2020-09-27 05:03:54,104] Trial 25 finished with value: 0.6894258959836473 and parameters: {'num_leaves': 28}. Best is trial 22 with value: 0.689088142033773.
num_leaves, val_score: 0.689088: 95%|#########5| 19/20 [00:19<00:00, 1.05it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000884 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.609273 valid's binary_logloss: 0.693331
Early stopping, best iteration is:
[34] train's binary_logloss: 0.656448 valid's binary_logloss: 0.689954
num_leaves, val_score: 0.689088: 100%|##########| 20/20 [00:20<00:00, 1.02it/s][I 2020-09-27 05:03:55,150] Trial 26 finished with value: 0.6899541998715152 and parameters: {'num_leaves': 117}. Best is trial 22 with value: 0.689088142033773.
num_leaves, val_score: 0.689088: 100%|##########| 20/20 [00:20<00:00, 1.04s/it]
bagging, val_score: 0.689088: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000959 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685151 valid's binary_logloss: 0.689666
[200] train's binary_logloss: 0.681327 valid's binary_logloss: 0.689749
Early stopping, best iteration is:
[143] train's binary_logloss: 0.68343 valid's binary_logloss: 0.689482
bagging, val_score: 0.689088: 10%|# | 1/10 [00:01<00:09, 1.05s/it][I 2020-09-27 05:03:56,212] Trial 27 finished with value: 0.6894817163491044 and parameters: {'bagging_fraction': 0.9223383259961062, 'bagging_freq': 3}. Best is trial 27 with value: 0.6894817163491044.
bagging, val_score: 0.689088: 10%|# | 1/10 [00:01<00:09, 1.05s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000943 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685286 valid's binary_logloss: 0.689382
[200] train's binary_logloss: 0.68151 valid's binary_logloss: 0.689555
Early stopping, best iteration is:
[173] train's binary_logloss: 0.682462 valid's binary_logloss: 0.688945
bagging, val_score: 0.688945: 20%|## | 2/10 [00:02<00:08, 1.04s/it][I 2020-09-27 05:03:57,220] Trial 28 finished with value: 0.6889445920177824 and parameters: {'bagging_fraction': 0.40058272142371876, 'bagging_freq': 7}. Best is trial 28 with value: 0.6889445920177824.
bagging, val_score: 0.688945: 20%|## | 2/10 [00:02<00:08, 1.04s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000924 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685224 valid's binary_logloss: 0.689275
[200] train's binary_logloss: 0.681491 valid's binary_logloss: 0.689706
Early stopping, best iteration is:
[168] train's binary_logloss: 0.682543 valid's binary_logloss: 0.688999
bagging, val_score: 0.688945: 30%|### | 3/10 [00:03<00:07, 1.03s/it][I 2020-09-27 05:03:58,234] Trial 29 finished with value: 0.6889992638135699 and parameters: {'bagging_fraction': 0.4181787978927057, 'bagging_freq': 7}. Best is trial 28 with value: 0.6889445920177824.
bagging, val_score: 0.688945: 30%|### | 3/10 [00:03<00:07, 1.03s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000991 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685284 valid's binary_logloss: 0.68908
[200] train's binary_logloss: 0.681564 valid's binary_logloss: 0.689249
Early stopping, best iteration is:
[172] train's binary_logloss: 0.682574 valid's binary_logloss: 0.688831
bagging, val_score: 0.688831: 40%|#### | 4/10 [00:04<00:06, 1.02s/it][I 2020-09-27 05:03:59,225] Trial 30 finished with value: 0.688830953603927 and parameters: {'bagging_fraction': 0.40441196806803215, 'bagging_freq': 7}. Best is trial 30 with value: 0.688830953603927.
bagging, val_score: 0.688831: 40%|#### | 4/10 [00:04<00:06, 1.02s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001020 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685401 valid's binary_logloss: 0.689413
[200] train's binary_logloss: 0.68163 valid's binary_logloss: 0.6894
Early stopping, best iteration is:
[172] train's binary_logloss: 0.682614 valid's binary_logloss: 0.688975
bagging, val_score: 0.688831: 50%|##### | 5/10 [00:05<00:05, 1.01s/it][I 2020-09-27 05:04:00,231] Trial 31 finished with value: 0.688974864123655 and parameters: {'bagging_fraction': 0.4053863270403766, 'bagging_freq': 7}. Best is trial 30 with value: 0.688830953603927.
bagging, val_score: 0.688831: 50%|##### | 5/10 [00:05<00:05, 1.01s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000972 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685259 valid's binary_logloss: 0.68914
Early stopping, best iteration is:
[95] train's binary_logloss: 0.685454 valid's binary_logloss: 0.689029
bagging, val_score: 0.688831: 60%|###### | 6/10 [00:05<00:03, 1.07it/s][I 2020-09-27 05:04:00,968] Trial 32 finished with value: 0.6890288131482014 and parameters: {'bagging_fraction': 0.4233037496739684, 'bagging_freq': 7}. Best is trial 30 with value: 0.688830953603927.
bagging, val_score: 0.688831: 60%|###### | 6/10 [00:05<00:03, 1.07it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000937 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685245 valid's binary_logloss: 0.689157
Early stopping, best iteration is:
[87] train's binary_logloss: 0.685847 valid's binary_logloss: 0.688973
bagging, val_score: 0.688831: 70%|####### | 7/10 [00:06<00:02, 1.16it/s][I 2020-09-27 05:04:01,669] Trial 33 finished with value: 0.6889732179547475 and parameters: {'bagging_fraction': 0.4054479427416367, 'bagging_freq': 7}. Best is trial 30 with value: 0.688830953603927.
bagging, val_score: 0.688831: 70%|####### | 7/10 [00:06<00:02, 1.16it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001013 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685369 valid's binary_logloss: 0.688941
[200] train's binary_logloss: 0.681625 valid's binary_logloss: 0.689259
Early stopping, best iteration is:
[174] train's binary_logloss: 0.682563 valid's binary_logloss: 0.688665
bagging, val_score: 0.688665: 80%|######## | 8/10 [00:07<00:01, 1.12it/s][I 2020-09-27 05:04:02,643] Trial 34 finished with value: 0.6886646750149633 and parameters: {'bagging_fraction': 0.4052909923895454, 'bagging_freq': 7}. Best is trial 34 with value: 0.6886646750149633.
bagging, val_score: 0.688665: 80%|######## | 8/10 [00:07<00:01, 1.12it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001120 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68534 valid's binary_logloss: 0.68936
[200] train's binary_logloss: 0.681589 valid's binary_logloss: 0.689277
Early stopping, best iteration is:
[160] train's binary_logloss: 0.682993 valid's binary_logloss: 0.689007
bagging, val_score: 0.688665: 90%|######### | 9/10 [00:08<00:00, 1.09it/s][I 2020-09-27 05:04:03,602] Trial 35 finished with value: 0.6890068245461886 and parameters: {'bagging_fraction': 0.40351986086829467, 'bagging_freq': 7}. Best is trial 34 with value: 0.6886646750149633.
bagging, val_score: 0.688665: 90%|######### | 9/10 [00:08<00:00, 1.09it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000983 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68515 valid's binary_logloss: 0.68937
[200] train's binary_logloss: 0.681294 valid's binary_logloss: 0.689514
Early stopping, best iteration is:
[113] train's binary_logloss: 0.684606 valid's binary_logloss: 0.68927
bagging, val_score: 0.688665: 100%|##########| 10/10 [00:09<00:00, 1.11it/s][I 2020-09-27 05:04:04,460] Trial 36 finished with value: 0.6892703778547441 and parameters: {'bagging_fraction': 0.5684326008067807, 'bagging_freq': 5}. Best is trial 34 with value: 0.6886646750149633.
bagging, val_score: 0.688665: 100%|##########| 10/10 [00:09<00:00, 1.08it/s]
feature_fraction_stage2, val_score: 0.688665: 0%| | 0/6 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000961 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685369 valid's binary_logloss: 0.688941
[200] train's binary_logloss: 0.681625 valid's binary_logloss: 0.689259
Early stopping, best iteration is:
[174] train's binary_logloss: 0.682563 valid's binary_logloss: 0.688665
feature_fraction_stage2, val_score: 0.688665: 17%|#6 | 1/6 [00:01<00:05, 1.01s/it][I 2020-09-27 05:04:05,490] Trial 37 finished with value: 0.6886646750149633 and parameters: {'feature_fraction': 0.516}. Best is trial 37 with value: 0.6886646750149633.
feature_fraction_stage2, val_score: 0.688665: 17%|#6 | 1/6 [00:01<00:05, 1.01s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000867 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685291 valid's binary_logloss: 0.689179
Early stopping, best iteration is:
[86] train's binary_logloss: 0.685898 valid's binary_logloss: 0.688923
feature_fraction_stage2, val_score: 0.688665: 33%|###3 | 2/6 [00:01<00:03, 1.09it/s][I 2020-09-27 05:04:06,195] Trial 38 finished with value: 0.6889230860549193 and parameters: {'feature_fraction': 0.45199999999999996}. Best is trial 37 with value: 0.6886646750149633.
feature_fraction_stage2, val_score: 0.688665: 33%|###3 | 2/6 [00:01<00:03, 1.09it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001184 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685287 valid's binary_logloss: 0.689043
[200] train's binary_logloss: 0.681498 valid's binary_logloss: 0.689164
Early stopping, best iteration is:
[168] train's binary_logloss: 0.682602 valid's binary_logloss: 0.688738
feature_fraction_stage2, val_score: 0.688665: 50%|##### | 3/6 [00:02<00:02, 1.07it/s][I 2020-09-27 05:04:07,161] Trial 39 finished with value: 0.6887381070029943 and parameters: {'feature_fraction': 0.5479999999999999}. Best is trial 37 with value: 0.6886646750149633.
feature_fraction_stage2, val_score: 0.688665: 50%|##### | 3/6 [00:02<00:02, 1.07it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000856 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685333 valid's binary_logloss: 0.689331
[200] train's binary_logloss: 0.681639 valid's binary_logloss: 0.689698
Early stopping, best iteration is:
[166] train's binary_logloss: 0.682754 valid's binary_logloss: 0.689155
feature_fraction_stage2, val_score: 0.688665: 67%|######6 | 4/6 [00:03<00:01, 1.07it/s][I 2020-09-27 05:04:08,093] Trial 40 finished with value: 0.6891552510228472 and parameters: {'feature_fraction': 0.42}. Best is trial 37 with value: 0.6886646750149633.
feature_fraction_stage2, val_score: 0.688665: 67%|######6 | 4/6 [00:03<00:01, 1.07it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001100 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685198 valid's binary_logloss: 0.689125
[200] train's binary_logloss: 0.681316 valid's binary_logloss: 0.689341
Early stopping, best iteration is:
[174] train's binary_logloss: 0.682274 valid's binary_logloss: 0.688933
feature_fraction_stage2, val_score: 0.688665: 83%|########3 | 5/6 [00:04<00:00, 1.04it/s][I 2020-09-27 05:04:09,134] Trial 41 finished with value: 0.6889333127717522 and parameters: {'feature_fraction': 0.58}. Best is trial 37 with value: 0.6886646750149633.
feature_fraction_stage2, val_score: 0.688665: 83%|########3 | 5/6 [00:04<00:00, 1.04it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011122 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685369 valid's binary_logloss: 0.688941
[200] train's binary_logloss: 0.681625 valid's binary_logloss: 0.689259
Early stopping, best iteration is:
[174] train's binary_logloss: 0.682563 valid's binary_logloss: 0.688665
feature_fraction_stage2, val_score: 0.688665: 100%|##########| 6/6 [00:05<00:00, 1.03it/s][I 2020-09-27 05:04:10,124] Trial 42 finished with value: 0.6886646750149633 and parameters: {'feature_fraction': 0.484}. Best is trial 37 with value: 0.6886646750149633.
feature_fraction_stage2, val_score: 0.688665: 100%|##########| 6/6 [00:05<00:00, 1.06it/s]
regularization_factors, val_score: 0.688665: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000996 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685322 valid's binary_logloss: 0.689114
[200] train's binary_logloss: 0.681508 valid's binary_logloss: 0.689186
Early stopping, best iteration is:
[175] train's binary_logloss: 0.682376 valid's binary_logloss: 0.688631
regularization_factors, val_score: 0.688631: 5%|5 | 1/20 [00:01<00:20, 1.08s/it][I 2020-09-27 05:04:11,228] Trial 43 finished with value: 0.6886306062996863 and parameters: {'lambda_l1': 0.05593044522293832, 'lambda_l2': 0.003082206943612091}. Best is trial 43 with value: 0.6886306062996863.
regularization_factors, val_score: 0.688631: 5%|5 | 1/20 [00:01<00:20, 1.08s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000926 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685295 valid's binary_logloss: 0.689098
Early stopping, best iteration is:
[86] train's binary_logloss: 0.68591 valid's binary_logloss: 0.688875
regularization_factors, val_score: 0.688631: 10%|# | 2/20 [00:01<00:17, 1.01it/s][I 2020-09-27 05:04:11,991] Trial 44 finished with value: 0.6888752025477913 and parameters: {'lambda_l1': 0.1570663292442244, 'lambda_l2': 0.0010606198676237839}. Best is trial 43 with value: 0.6886306062996863.
regularization_factors, val_score: 0.688631: 10%|# | 2/20 [00:01<00:17, 1.01it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000960 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.68561 valid's binary_logloss: 0.689202
Early stopping, best iteration is:
[89] train's binary_logloss: 0.686012 valid's binary_logloss: 0.688996
regularization_factors, val_score: 0.688631: 15%|#5 | 3/20 [00:02<00:15, 1.10it/s][I 2020-09-27 05:04:12,715] Trial 45 finished with value: 0.6889959749240171 and parameters: {'lambda_l1': 1.4927818545611396e-06, 'lambda_l2': 7.613862377908525}. Best is trial 43 with value: 0.6886306062996863.
regularization_factors, val_score: 0.688631: 15%|#5 | 3/20 [00:02<00:15, 1.10it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000944 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685965 valid's binary_logloss: 0.689511
[200] train's binary_logloss: 0.682794 valid's binary_logloss: 0.689793
Early stopping, best iteration is:
[174] train's binary_logloss: 0.683538 valid's binary_logloss: 0.689222
regularization_factors, val_score: 0.688631: 20%|## | 4/20 [00:03<00:15, 1.05it/s][I 2020-09-27 05:04:13,760] Trial 46 finished with value: 0.6892218212122403 and parameters: {'lambda_l1': 5.48610084980033, 'lambda_l2': 0.00024090694580484778}. Best is trial 43 with value: 0.6886306062996863.
regularization_factors, val_score: 0.688631: 20%|## | 4/20 [00:03<00:15, 1.05it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.004926 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685369 valid's binary_logloss: 0.688941
[200] train's binary_logloss: 0.681626 valid's binary_logloss: 0.689259
Early stopping, best iteration is:
[174] train's binary_logloss: 0.682563 valid's binary_logloss: 0.688665
regularization_factors, val_score: 0.688631: 25%|##5 | 5/20 [00:04<00:14, 1.02it/s][I 2020-09-27 05:04:14,825] Trial 47 finished with value: 0.6886646716779263 and parameters: {'lambda_l1': 0.0011713348102023722, 'lambda_l2': 1.4326223733321923e-08}. Best is trial 43 with value: 0.6886306062996863.
regularization_factors, val_score: 0.688631: 25%|##5 | 5/20 [00:04<00:14, 1.02it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001038 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685369 valid's binary_logloss: 0.688941
[200] train's binary_logloss: 0.681626 valid's binary_logloss: 0.689259
Early stopping, best iteration is:
[174] train's binary_logloss: 0.682563 valid's binary_logloss: 0.688665
regularization_factors, val_score: 0.688631: 30%|### | 6/20 [00:05<00:14, 1.07s/it][I 2020-09-27 05:04:16,093] Trial 48 finished with value: 0.6886646717918603 and parameters: {'lambda_l1': 0.0011323335612472105, 'lambda_l2': 1.2449590491320676e-08}. Best is trial 43 with value: 0.6886306062996863.
regularization_factors, val_score: 0.688631: 30%|### | 6/20 [00:05<00:14, 1.07s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009231 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685369 valid's binary_logloss: 0.688941
[200] train's binary_logloss: 0.681626 valid's binary_logloss: 0.689259
Early stopping, best iteration is:
[174] train's binary_logloss: 0.682563 valid's binary_logloss: 0.688665
regularization_factors, val_score: 0.688631: 35%|###5 | 7/20 [00:06<00:13, 1.05s/it][I 2020-09-27 05:04:17,084] Trial 49 finished with value: 0.6886646722615538 and parameters: {'lambda_l1': 0.0009667778650190732, 'lambda_l2': 1.4964282342181056e-08}. Best is trial 43 with value: 0.6886306062996863.
regularization_factors, val_score: 0.688631: 35%|###5 | 7/20 [00:06<00:13, 1.05s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001006 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685369 valid's binary_logloss: 0.688941
[200] train's binary_logloss: 0.681626 valid's binary_logloss: 0.689259
Early stopping, best iteration is:
[174] train's binary_logloss: 0.682563 valid's binary_logloss: 0.688665
regularization_factors, val_score: 0.688631: 40%|#### | 8/20 [00:07<00:12, 1.04s/it][I 2020-09-27 05:04:18,110] Trial 50 finished with value: 0.6886646726282439 and parameters: {'lambda_l1': 0.000834005476706058, 'lambda_l2': 1.1339953988984425e-08}. Best is trial 43 with value: 0.6886306062996863.
regularization_factors, val_score: 0.688631: 40%|#### | 8/20 [00:07<00:12, 1.04s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001002 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685369 valid's binary_logloss: 0.688941
[200] train's binary_logloss: 0.681626 valid's binary_logloss: 0.689259
Early stopping, best iteration is:
[174] train's binary_logloss: 0.682563 valid's binary_logloss: 0.688665
regularization_factors, val_score: 0.688631: 45%|####5 | 9/20 [00:09<00:11, 1.04s/it][I 2020-09-27 05:04:19,146] Trial 51 finished with value: 0.6886646724473121 and parameters: {'lambda_l1': 0.0008969854861021092, 'lambda_l2': 1.1117633466286682e-08}. Best is trial 43 with value: 0.6886306062996863.
regularization_factors, val_score: 0.688631: 45%|####5 | 9/20 [00:09<00:11, 1.04s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013191 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685369 valid's binary_logloss: 0.688941
[200] train's binary_logloss: 0.681626 valid's binary_logloss: 0.689259
Early stopping, best iteration is:
[174] train's binary_logloss: 0.682563 valid's binary_logloss: 0.688665
regularization_factors, val_score: 0.688631: 50%|##### | 10/20 [00:09<00:10, 1.01s/it][I 2020-09-27 05:04:20,098] Trial 52 finished with value: 0.6886646724367362 and parameters: {'lambda_l1': 0.0009016551190739346, 'lambda_l2': 1.0925317395658214e-08}. Best is trial 43 with value: 0.6886306062996863.
regularization_factors, val_score: 0.688631: 50%|##### | 10/20 [00:09<00:10, 1.01s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000952 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685369 valid's binary_logloss: 0.688941
[200] train's binary_logloss: 0.681626 valid's binary_logloss: 0.689259
Early stopping, best iteration is:
[174] train's binary_logloss: 0.682563 valid's binary_logloss: 0.688665
regularization_factors, val_score: 0.688631: 55%|#####5 | 11/20 [00:10<00:09, 1.01s/it][I 2020-09-27 05:04:21,108] Trial 53 finished with value: 0.6886646727381388 and parameters: {'lambda_l1': 0.0007965344637915165, 'lambda_l2': 1.3659640884572439e-08}. Best is trial 43 with value: 0.6886306062996863.
regularization_factors, val_score: 0.688631: 55%|#####5 | 11/20 [00:10<00:09, 1.01s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011699 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685369 valid's binary_logloss: 0.688941
[200] train's binary_logloss: 0.681626 valid's binary_logloss: 0.689259
Early stopping, best iteration is:
[174] train's binary_logloss: 0.682563 valid's binary_logloss: 0.688665
regularization_factors, val_score: 0.688631: 60%|###### | 12/20 [00:11<00:07, 1.01it/s][I 2020-09-27 05:04:22,056] Trial 54 finished with value: 0.6886646716630922 and parameters: {'lambda_l1': 0.0011776384915221255, 'lambda_l2': 1.0468618620438098e-08}. Best is trial 43 with value: 0.6886306062996863.
regularization_factors, val_score: 0.688631: 60%|###### | 12/20 [00:11<00:07, 1.01it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000972 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685369 valid's binary_logloss: 0.688941
[200] train's binary_logloss: 0.681589 valid's binary_logloss: 0.689058
Early stopping, best iteration is:
[174] train's binary_logloss: 0.68257 valid's binary_logloss: 0.688662
regularization_factors, val_score: 0.688631: 65%|######5 | 13/20 [00:12<00:06, 1.00it/s][I 2020-09-27 05:04:23,068] Trial 55 finished with value: 0.6886618433996927 and parameters: {'lambda_l1': 0.00344335006832809, 'lambda_l2': 1.0707613854300203e-08}. Best is trial 43 with value: 0.6886306062996863.
regularization_factors, val_score: 0.688631: 65%|######5 | 13/20 [00:12<00:06, 1.00it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000963 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685283 valid's binary_logloss: 0.689021
[200] train's binary_logloss: 0.68165 valid's binary_logloss: 0.688997
Early stopping, best iteration is:
[172] train's binary_logloss: 0.682609 valid's binary_logloss: 0.688401
regularization_factors, val_score: 0.688401: 70%|####### | 14/20 [00:13<00:05, 1.00it/s][I 2020-09-27 05:04:24,069] Trial 56 finished with value: 0.6884008811372169 and parameters: {'lambda_l1': 0.013564067517829166, 'lambda_l2': 0.0028990672498837708}. Best is trial 56 with value: 0.6884008811372169.
regularization_factors, val_score: 0.688401: 70%|####### | 14/20 [00:13<00:05, 1.00it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001378 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685305 valid's binary_logloss: 0.689126
[200] train's binary_logloss: 0.681644 valid's binary_logloss: 0.689459
Early stopping, best iteration is:
[187] train's binary_logloss: 0.68211 valid's binary_logloss: 0.688878
regularization_factors, val_score: 0.688401: 75%|#######5 | 15/20 [00:14<00:05, 1.02s/it][I 2020-09-27 05:04:25,134] Trial 57 finished with value: 0.6888781933882588 and parameters: {'lambda_l1': 0.06847066607771893, 'lambda_l2': 0.005100079808171161}. Best is trial 56 with value: 0.6884008811372169.
regularization_factors, val_score: 0.688401: 75%|#######5 | 15/20 [00:14<00:05, 1.02s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000979 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685375 valid's binary_logloss: 0.689304
[200] train's binary_logloss: 0.681532 valid's binary_logloss: 0.689352
Early stopping, best iteration is:
[173] train's binary_logloss: 0.682496 valid's binary_logloss: 0.688928
regularization_factors, val_score: 0.688401: 80%|######## | 16/20 [00:15<00:04, 1.01s/it][I 2020-09-27 05:04:26,139] Trial 58 finished with value: 0.6889282775916249 and parameters: {'lambda_l1': 0.04206907981702274, 'lambda_l2': 0.008049472807870911}. Best is trial 56 with value: 0.6884008811372169.
regularization_factors, val_score: 0.688401: 80%|######## | 16/20 [00:15<00:04, 1.01s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001050 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685283 valid's binary_logloss: 0.689021
[200] train's binary_logloss: 0.68165 valid's binary_logloss: 0.688997
Early stopping, best iteration is:
[172] train's binary_logloss: 0.682609 valid's binary_logloss: 0.688401
regularization_factors, val_score: 0.688401: 85%|########5 | 17/20 [00:19<00:05, 1.68s/it][I 2020-09-27 05:04:29,386] Trial 59 finished with value: 0.6884009077923067 and parameters: {'lambda_l1': 0.01818292826675875, 'lambda_l2': 4.793765360259705e-07}. Best is trial 56 with value: 0.6884008811372169.
regularization_factors, val_score: 0.688401: 85%|########5 | 17/20 [00:19<00:05, 1.68s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001800 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685284 valid's binary_logloss: 0.689021
[200] train's binary_logloss: 0.681652 valid's binary_logloss: 0.688997
Early stopping, best iteration is:
[172] train's binary_logloss: 0.68261 valid's binary_logloss: 0.688401
regularization_factors, val_score: 0.688401: 90%|######### | 18/20 [00:20<00:03, 1.51s/it][I 2020-09-27 05:04:30,486] Trial 60 finished with value: 0.6884009815019274 and parameters: {'lambda_l1': 0.022406843035860227, 'lambda_l2': 2.0488930606948696e-06}. Best is trial 56 with value: 0.6884008811372169.
regularization_factors, val_score: 0.688401: 90%|######### | 18/20 [00:20<00:03, 1.51s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001595 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685284 valid's binary_logloss: 0.689021
[200] train's binary_logloss: 0.681652 valid's binary_logloss: 0.688997
Early stopping, best iteration is:
[172] train's binary_logloss: 0.682611 valid's binary_logloss: 0.688401
regularization_factors, val_score: 0.688401: 95%|#########5| 19/20 [00:21<00:01, 1.39s/it][I 2020-09-27 05:04:31,591] Trial 61 finished with value: 0.6884010228456562 and parameters: {'lambda_l1': 0.024768099089603052, 'lambda_l2': 5.962389419495749e-06}. Best is trial 56 with value: 0.6884008811372169.
regularization_factors, val_score: 0.688401: 95%|#########5| 19/20 [00:21<00:01, 1.39s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001134 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685378 valid's binary_logloss: 0.688831
[200] train's binary_logloss: 0.681663 valid's binary_logloss: 0.688838
Early stopping, best iteration is:
[182] train's binary_logloss: 0.682288 valid's binary_logloss: 0.688653
regularization_factors, val_score: 0.688401: 100%|##########| 20/20 [00:22<00:00, 1.32s/it][I 2020-09-27 05:04:32,742] Trial 62 finished with value: 0.6886532224197758 and parameters: {'lambda_l1': 0.03746482418394038, 'lambda_l2': 1.895168169019993e-06}. Best is trial 56 with value: 0.6884008811372169.
regularization_factors, val_score: 0.688401: 100%|##########| 20/20 [00:22<00:00, 1.13s/it]
min_data_in_leaf, val_score: 0.688401: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000979 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685537 valid's binary_logloss: 0.689132
Early stopping, best iteration is:
[95] train's binary_logloss: 0.68572 valid's binary_logloss: 0.689056
min_data_in_leaf, val_score: 0.688401: 20%|## | 1/5 [00:00<00:03, 1.30it/s][I 2020-09-27 05:04:33,525] Trial 63 finished with value: 0.6890556326625765 and parameters: {'min_child_samples': 100}. Best is trial 63 with value: 0.6890556326625765.
min_data_in_leaf, val_score: 0.688401: 20%|## | 1/5 [00:00<00:03, 1.30it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000984 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685507 valid's binary_logloss: 0.688911
Early stopping, best iteration is:
[86] train's binary_logloss: 0.68607 valid's binary_logloss: 0.688678
min_data_in_leaf, val_score: 0.688401: 40%|#### | 2/5 [00:01<00:02, 1.32it/s][I 2020-09-27 05:04:34,254] Trial 64 finished with value: 0.68867811407117 and parameters: {'min_child_samples': 50}. Best is trial 64 with value: 0.68867811407117.
min_data_in_leaf, val_score: 0.688401: 40%|#### | 2/5 [00:01<00:02, 1.32it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001032 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685329 valid's binary_logloss: 0.688996
[200] train's binary_logloss: 0.681584 valid's binary_logloss: 0.689202
Early stopping, best iteration is:
[168] train's binary_logloss: 0.682673 valid's binary_logloss: 0.688789
min_data_in_leaf, val_score: 0.688401: 60%|###### | 3/5 [00:02<00:01, 1.20it/s][I 2020-09-27 05:04:35,256] Trial 65 finished with value: 0.6887885221160116 and parameters: {'min_child_samples': 25}. Best is trial 64 with value: 0.68867811407117.
min_data_in_leaf, val_score: 0.688401: 60%|###### | 3/5 [00:02<00:01, 1.20it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001002 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685286 valid's binary_logloss: 0.689236
Early stopping, best iteration is:
[87] train's binary_logloss: 0.685815 valid's binary_logloss: 0.68908
min_data_in_leaf, val_score: 0.688401: 80%|######## | 4/5 [00:03<00:00, 1.27it/s][I 2020-09-27 05:04:35,953] Trial 66 finished with value: 0.6890800984264625 and parameters: {'min_child_samples': 10}. Best is trial 64 with value: 0.68867811407117.
min_data_in_leaf, val_score: 0.688401: 80%|######## | 4/5 [00:03<00:00, 1.27it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663
[LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000956 seconds.
You can set `force_row_wise=true` to remove the overhead.
And if memory is not enough, you can set `force_col_wise=true`.
[LightGBM] [Info] Total Bins 4689
[LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26
[LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450
[LightGBM] [Info] Start training from score -0.006450
Training until validation scores don't improve for 100 rounds
[100] train's binary_logloss: 0.685281 valid's binary_logloss: 0.688938
Early stopping, best iteration is:
[96] train's binary_logloss: 0.685434 valid's binary_logloss: 0.688858
min_data_in_leaf, val_score: 0.688401: 100%|##########| 5/5 [00:03<00:00, 1.29it/s][I 2020-09-27 05:04:36,693] Trial 67 finished with value: 0.6888576677735015 and parameters: {'min_child_samples': 5}. Best is trial 64 with value: 0.68867811407117.
min_data_in_leaf, val_score: 0.688401: 100%|##########| 5/5 [00:03<00:00, 1.27it/s]
################################
CV_score:0.5369671988207919
---------------------------------
total CV_score:0.5760159134509492