t_mochizuki0
スプラ2をプレイした方はご存知の通り、このゲームには以下のモードがあります。
(今回のコンペに関係ないモードは除外)
各モードは勝利条件が異なるため、それぞれ別のゲームと言えるほど、プレーヤーの行動が変わります。
有利/不利なブキの組み合わせも、モードによって変わる可能性があります。
このため今回のデータの学習も、モード別に実施した方がより正確になるのでは? という動機です。
以下の学習結果を比較し、どちらの結果が上か確認します。
学習モデルは以下を使用させていただきました。
「LightGBMを使ったBase line」
なお、「2.」は各モード毎に学習データを分けて学習し、最後に結果をまとめています。(力技です)
「2.」の方が、Publicスコアが上がりました。
学習結果のCVを見ると、特に「ナワバリバトル」はCV=0.5996257033039711と高い値でした。
他のモードと分けた効果が出ていると思われます。
ただし、「4. 課題 (1)」で述べる通り、学習データに欠損値が生じるという問題があります。
このため上記は参考値と捉えてください。
全データをまとめて学習した際は欠損値が出なかったのですが、モード別に学習データを分けると欠損値が生じます。
change_to_target2()のforループ内で欠損値が発生するのですが、原因は未確認です。
何かお気づきの点がありましたら、コメントで教えていただけると助かります。
上記方法はナワバリバトルのCVは高かったものの、ガチエリア〜ガチヤグラのCVは低めでした(特にガチヤグラ)。
データ分割により、学習データ自体が減ってしまったことが原因かもしれません。
ガチエリア〜ガチヤグラは「ガチマッチ」としてまとめて学習します。これにより学習データ数を確保します。
ナワバリバトルはCVが良かったので、分けて学習します。
これらのモードは学習データの"lobby-mode"で判別できます。
以下の通り、全体としてはCVが上がりました。
ただしガチマッッチのCVはそれほど向上しなかったので、微妙なところかもしれません。
LightGBMのモデルを開示していただいたOreginさん、ありがとうございました。
基準となるモデルがあるとこのような比較がやりやすいため、たいへん助かりました。
# ライブラリのインポート
import pandas as pd
import numpy as np
import re
import matplotlib.pyplot as plt
import seaborn as sns
import lightgbm as lgb
from sklearn.model_selection import train_test_split
from sklearn.model_selection import KFold
from sklearn.metrics import accuracy_score
import warnings
warnings.filterwarnings('ignore')
# データの読込
train = pd.read_csv("../data/train_data.csv")
test = pd.read_csv('../data/test_data.csv')
# 欠損値は、全て「-1」とする。
def fill_all_null(df):
for col_name in df.columns[df.isnull().sum()!=0]:
df[col_name] = df[col_name].fillna(-1)
# 訓練データ、テストデータの欠損値を補完
fill_all_null(train)
fill_all_null(test)
# ターゲットエンコーディングの関数定義
def change_to_target2(train_df,test_df,input_column_name,output_column_name):
from sklearn.model_selection import KFold
# nan埋め処理
train[input_column_name] = train[input_column_name].fillna('-1').isnull().sum()
test[input_column_name] = test[input_column_name].fillna('-1').isnull().sum()
kf = KFold(n_splits=5, shuffle=True, random_state=71)
#=========================================================#
c=input_column_name
# 学習データ全体で各カテゴリにおけるyの平均を計算
data_tmp = pd.DataFrame({c: train_df[c],'target':train_df['y']})
target_mean = data_tmp.groupby(c)['target'].mean()
#テストデータのカテゴリを置換
test_df[output_column_name] = test_df[c].map(target_mean)
# 変換後の値を格納する配列を準備
tmp = np.repeat(np.nan, train_df.shape[0])
for i, (train_index, test_index) in enumerate(kf.split(train_df)): # NFOLDS回まわる
#学習データについて、各カテゴリにおける目的変数の平均を計算
target_mean = data_tmp.iloc[train_index].groupby(c)['target'].mean()
#バリデーションデータについて、変換後の値を一時配列に格納
tmp[test_index] = train_df[c].iloc[test_index].map(target_mean)
#変換後のデータで元の変数を置換
train_df[output_column_name] = tmp
#========================================================#
# 'period','game-ver','mode','lobby'の各列は、勝敗に大きく影響しないと思われるため削除
train = train.drop(labels=['period', 'game-ver', 'mode', 'lobby'], axis=1)
test = test.drop(labels=['period', 'game-ver', 'mode', 'lobby'], axis=1)
# regular, gachiの各lobby-mode別データを取得する
train_regular = train[train['lobby-mode'] == "regular"]
train_gachi = train[train['lobby-mode'] == "gachi"]
test_regular = test[test['lobby-mode'] == "regular"]
test_gachi = test[test['lobby-mode'] == "gachi"]
# オブジェクトの列のリストを作成
object_col_list = train.select_dtypes(include=object).columns
# オブジェクトの列は全てターゲットエンコーディング実施
for col in object_col_list:
change_to_target2(train_regular, test_regular, col, "enc_"+col)
change_to_target2(train_gachi, test_gachi, col, "enc_"+col)
# 変換前の列を削除
train_regular = train_regular.drop(object_col_list,axis=1)
train_gachi = train_gachi.drop(object_col_list,axis=1)
test_regular = test_regular.drop(object_col_list,axis=1)
test_gachi = test_gachi.drop(object_col_list,axis=1)
# 予測結果をあとでまとめるために、idの列を抽出する
test_regular_id = test_regular[['id']].copy() # DataFrameとして抽出するために、リストをネストする
test_gachi_id = test_gachi[['id']].copy()
# 'id'の列を削除
train_regular = train_regular.drop('id',axis=1)
train_gachi = train_gachi.drop('id',axis=1)
test_regular = test_regular.drop('id',axis=1)
test_gachi = test_gachi.drop('id',axis=1)
# 訓練データに欠損がないことの確認(欠損値がある項目数を表示)
print(train_regular.isnull().values.sum())
print(train_gachi.isnull().values.sum())
# テストデータに欠損がないことの確認(欠損値がある項目数を表示)
print(test_regular.isnull().values.sum())
print(test_gachi.isnull().values.sum())
2 0 0 2
# 訓練データを説明変数と目的変数に分割
target_regular = train_regular['y']
train_x_regular = train_regular.drop('y',axis=1)
target_gachi = train_gachi['y']
train_x_gachi = train_gachi.drop('y',axis=1)
# LGBMのパラメータを設定
params = {
# 二値分類問題
'objective': 'binary',
# 損失関数は二値のlogloss
#'metric': 'auc',
'metric': 'binary_logloss',
# 最大イテレーション回数指定
'num_iterations' : 1000,
# early_stopping 回数指定
'early_stopping_rounds' : 100,
}
def predict(train_x, target, test):
# k-分割交差検証を使って学習&予測(K=10)
FOLD_NUM = 10
kf = KFold(n_splits=FOLD_NUM,
random_state=42)
#検証時のスコアを初期化
scores = []
#テストデータの予測値を初期化
pred_cv = np.zeros(len(test.index))
#lgbmのラウンド数を定義
num_round = 10000
for i, (tdx, vdx) in enumerate(kf.split(train_x, target)):
print(f'Fold : {i}')
# 訓練用データと検証用データに分割
X_train, X_valid, y_train, y_valid = train_x.iloc[tdx], train_x.iloc[vdx], target.values[tdx], target.values[vdx]
lgb_train = lgb.Dataset(X_train, y_train)
lgb_valid = lgb.Dataset(X_valid, y_valid)
# 学習の実行
model = lgb.train(params, lgb_train, num_boost_round=num_round,
valid_names=["train", "valid"], valid_sets=[lgb_train, lgb_valid],
verbose_eval=100)
# 検証データに対する予測値を求めて、勝敗(0 or 1)に変換
va_pred = np.round(model.predict(X_valid,num_iteration=model.best_iteration))
# accuracyスコアを計算
score_ = accuracy_score(y_valid, va_pred)
# フォールド毎の検証時のスコアを格納
scores.append(score_)
#テストデータに対する予測値を求める
submission = model.predict(test,num_iteration=model.best_iteration)
#テストデータに対する予測値をフォールド数で割って蓄積
#(フォールド毎の予測値の平均値を求めることと同じ)
pred_cv += submission/FOLD_NUM
# 最終的なテストデータに対する予測値を勝敗(0 or 1)に変換
pred_cv = np.round(pred_cv)
# 最終的なaccuracyスコアを平均値で出力
print('')
print('################################')
print('CV_score:'+ str(np.mean(scores)))
print()
return(scores, pred_cv)
# 各モード毎のスコアを初期化
total_scores = []
# 各モードの予測を実行
results = predict(train_x_regular, target_regular, test_regular)
total_scores.append(results[0])
test_regular_id['y'] = results[1]
results = predict(train_x_gachi, target_gachi, test_gachi)
total_scores.append(results[0])
test_gachi_id['y'] = results[1]
# 最終的なaccuracyスコアを出力
print('---------------------------------')
print('total CV_score:'+ str(np.mean(total_scores)))
Fold : 0 [I 2020-09-27 04:38:19,970] A new study created in memory with name: no-name-4ce5d708-0a14-4cfc-bc23-b74b1ae8bc8d feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000784 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.574988 valid's binary_logloss: 0.650038 [200] train's binary_logloss: 0.51996 valid's binary_logloss: 0.651275 Early stopping, best iteration is: [137] train's binary_logloss: 0.552891 valid's binary_logloss: 0.649537 feature_fraction, val_score: 0.649537: 14%|#4 | 1/7 [00:00<00:04, 1.41it/s][I 2020-09-27 04:38:20,696] Trial 0 finished with value: 0.6495366523489564 and parameters: {'feature_fraction': 0.7}. Best is trial 0 with value: 0.6495366523489564. feature_fraction, val_score: 0.649537: 14%|#4 | 1/7 [00:00<00:04, 1.41it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000262 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.579394 valid's binary_logloss: 0.650899 Early stopping, best iteration is: [92] train's binary_logloss: 0.584248 valid's binary_logloss: 0.650087 feature_fraction, val_score: 0.649537: 29%|##8 | 2/7 [00:01<00:03, 1.58it/s][I 2020-09-27 04:38:21,152] Trial 1 finished with value: 0.6500867396462259 and parameters: {'feature_fraction': 0.5}. Best is trial 0 with value: 0.6495366523489564. feature_fraction, val_score: 0.649537: 29%|##8 | 2/7 [00:01<00:03, 1.58it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000583 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.583335 valid's binary_logloss: 0.652707 Early stopping, best iteration is: [88] train's binary_logloss: 0.590815 valid's binary_logloss: 0.652179 feature_fraction, val_score: 0.649537: 43%|####2 | 3/7 [00:01<00:02, 1.77it/s][I 2020-09-27 04:38:21,556] Trial 2 finished with value: 0.652179314698687 and parameters: {'feature_fraction': 0.4}. Best is trial 0 with value: 0.6495366523489564. feature_fraction, val_score: 0.649537: 43%|####2 | 3/7 [00:01<00:02, 1.77it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000873 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.573386 valid's binary_logloss: 0.65332 Early stopping, best iteration is: [72] train's binary_logloss: 0.592163 valid's binary_logloss: 0.652672 feature_fraction, val_score: 0.649537: 57%|#####7 | 4/7 [00:02<00:01, 1.90it/s][I 2020-09-27 04:38:21,994] Trial 3 finished with value: 0.6526718623826693 and parameters: {'feature_fraction': 0.8}. Best is trial 0 with value: 0.6495366523489564. feature_fraction, val_score: 0.649537: 57%|#####7 | 4/7 [00:02<00:01, 1.90it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000825 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.573309 valid's binary_logloss: 0.653575 Early stopping, best iteration is: [67] train's binary_logloss: 0.59526 valid's binary_logloss: 0.652553 feature_fraction, val_score: 0.649537: 71%|#######1 | 5/7 [00:02<00:01, 1.58it/s][I 2020-09-27 04:38:22,880] Trial 4 finished with value: 0.6525525222524156 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 0 with value: 0.6495366523489564. feature_fraction, val_score: 0.649537: 71%|#######1 | 5/7 [00:02<00:01, 1.58it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.006985 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.576836 valid's binary_logloss: 0.653247 Early stopping, best iteration is: [70] train's binary_logloss: 0.596318 valid's binary_logloss: 0.651989 feature_fraction, val_score: 0.649537: 86%|########5 | 6/7 [00:03<00:00, 1.79it/s][I 2020-09-27 04:38:23,260] Trial 5 finished with value: 0.6519889736642206 and parameters: {'feature_fraction': 0.6}. Best is trial 0 with value: 0.6495366523489564. feature_fraction, val_score: 0.649537: 86%|########5 | 6/7 [00:03<00:00, 1.79it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000890 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.571148 valid's binary_logloss: 0.654809 Early stopping, best iteration is: [46] train's binary_logloss: 0.611505 valid's binary_logloss: 0.653731 feature_fraction, val_score: 0.649537: 100%|##########| 7/7 [00:03<00:00, 1.96it/s][I 2020-09-27 04:38:23,655] Trial 6 finished with value: 0.6537305766184812 and parameters: {'feature_fraction': 1.0}. Best is trial 0 with value: 0.6495366523489564. feature_fraction, val_score: 0.649537: 100%|##########| 7/7 [00:03<00:00, 1.91it/s] num_leaves, val_score: 0.649537: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000620 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.57822 valid's binary_logloss: 0.649469 Early stopping, best iteration is: [85] train's binary_logloss: 0.587383 valid's binary_logloss: 0.649227 num_leaves, val_score: 0.649227: 5%|5 | 1/20 [00:00<00:09, 2.10it/s][I 2020-09-27 04:38:24,142] Trial 7 finished with value: 0.6492270930479551 and parameters: {'num_leaves': 30}. Best is trial 7 with value: 0.6492270930479551. num_leaves, val_score: 0.649227: 5%|5 | 1/20 [00:00<00:09, 2.10it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000836 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.514021 valid's binary_logloss: 0.655753 Early stopping, best iteration is: [46] train's binary_logloss: 0.579148 valid's binary_logloss: 0.653741 num_leaves, val_score: 0.649227: 10%|# | 2/20 [00:01<00:08, 2.04it/s][I 2020-09-27 04:38:24,666] Trial 8 finished with value: 0.6537410492395737 and parameters: {'num_leaves': 60}. Best is trial 7 with value: 0.6492270930479551. num_leaves, val_score: 0.649227: 10%|# | 2/20 [00:01<00:08, 2.04it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.003688 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.635569 valid's binary_logloss: 0.650064 [200] train's binary_logloss: 0.617663 valid's binary_logloss: 0.647802 [300] train's binary_logloss: 0.603448 valid's binary_logloss: 0.648074 Early stopping, best iteration is: [204] train's binary_logloss: 0.617119 valid's binary_logloss: 0.647726 num_leaves, val_score: 0.647726: 15%|#5 | 3/20 [00:01<00:08, 2.08it/s][I 2020-09-27 04:38:25,128] Trial 9 finished with value: 0.6477259914776615 and parameters: {'num_leaves': 8}. Best is trial 9 with value: 0.6477259914776615. num_leaves, val_score: 0.647726: 15%|#5 | 3/20 [00:01<00:08, 2.08it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.003613 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.295966 valid's binary_logloss: 0.669876 Early stopping, best iteration is: [28] train's binary_logloss: 0.505319 valid's binary_logloss: 0.658492 num_leaves, val_score: 0.647726: 20%|## | 4/20 [00:02<00:11, 1.33it/s][I 2020-09-27 04:38:26,505] Trial 10 finished with value: 0.6584916550281572 and parameters: {'num_leaves': 229}. Best is trial 9 with value: 0.6477259914776615. num_leaves, val_score: 0.647726: 20%|## | 4/20 [00:02<00:11, 1.33it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.003890 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.614941 valid's binary_logloss: 0.650113 [200] train's binary_logloss: 0.584546 valid's binary_logloss: 0.649317 [300] train's binary_logloss: 0.558777 valid's binary_logloss: 0.650116 Early stopping, best iteration is: [244] train's binary_logloss: 0.57318 valid's binary_logloss: 0.649088 num_leaves, val_score: 0.647726: 25%|##5 | 5/20 [00:03<00:11, 1.29it/s][I 2020-09-27 04:38:27,335] Trial 11 finished with value: 0.6490878119710577 and parameters: {'num_leaves': 15}. Best is trial 9 with value: 0.6477259914776615. num_leaves, val_score: 0.647726: 25%|##5 | 5/20 [00:03<00:11, 1.29it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000790 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.646526 valid's binary_logloss: 0.653926 [200] train's binary_logloss: 0.634088 valid's binary_logloss: 0.649382 [300] train's binary_logloss: 0.62524 valid's binary_logloss: 0.648291 [400] train's binary_logloss: 0.617685 valid's binary_logloss: 0.648421 Early stopping, best iteration is: [324] train's binary_logloss: 0.623449 valid's binary_logloss: 0.648061 num_leaves, val_score: 0.647726: 30%|### | 6/20 [00:04<00:10, 1.34it/s][I 2020-09-27 04:38:28,014] Trial 12 finished with value: 0.6480614052342452 and parameters: {'num_leaves': 5}. Best is trial 9 with value: 0.6477259914776615. num_leaves, val_score: 0.647726: 30%|### | 6/20 [00:04<00:10, 1.34it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008792 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.414675 valid's binary_logloss: 0.664247 Early stopping, best iteration is: [26] train's binary_logloss: 0.571385 valid's binary_logloss: 0.659552 num_leaves, val_score: 0.647726: 35%|###5 | 7/20 [00:05<00:09, 1.37it/s][I 2020-09-27 04:38:28,706] Trial 13 finished with value: 0.6595522413168151 and parameters: {'num_leaves': 122}. Best is trial 9 with value: 0.6477259914776615. num_leaves, val_score: 0.647726: 35%|###5 | 7/20 [00:05<00:09, 1.37it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000826 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.470281 valid's binary_logloss: 0.657387 Early stopping, best iteration is: [37] train's binary_logloss: 0.570263 valid's binary_logloss: 0.654921 num_leaves, val_score: 0.647726: 40%|#### | 8/20 [00:05<00:08, 1.36it/s][I 2020-09-27 04:38:29,454] Trial 14 finished with value: 0.654921475007968 and parameters: {'num_leaves': 85}. Best is trial 9 with value: 0.6477259914776615. num_leaves, val_score: 0.647726: 40%|#### | 8/20 [00:05<00:08, 1.36it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004565 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.635569 valid's binary_logloss: 0.650064 [200] train's binary_logloss: 0.617663 valid's binary_logloss: 0.647802 [300] train's binary_logloss: 0.603448 valid's binary_logloss: 0.648074 Early stopping, best iteration is: [204] train's binary_logloss: 0.617119 valid's binary_logloss: 0.647726 num_leaves, val_score: 0.647726: 45%|####5 | 9/20 [00:06<00:08, 1.24it/s][I 2020-09-27 04:38:30,428] Trial 15 finished with value: 0.6477259914776615 and parameters: {'num_leaves': 8}. Best is trial 9 with value: 0.6477259914776615. num_leaves, val_score: 0.647726: 45%|####5 | 9/20 [00:06<00:08, 1.24it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000847 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.291012 valid's binary_logloss: 0.673269 Early stopping, best iteration is: [36] train's binary_logloss: 0.469537 valid's binary_logloss: 0.661474 num_leaves, val_score: 0.647726: 50%|##### | 10/20 [00:07<00:09, 1.07it/s][I 2020-09-27 04:38:31,655] Trial 16 finished with value: 0.661474106398792 and parameters: {'num_leaves': 233}. Best is trial 9 with value: 0.6477259914776615. num_leaves, val_score: 0.647726: 50%|##### | 10/20 [00:07<00:09, 1.07it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007106 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.35312 valid's binary_logloss: 0.66936 Early stopping, best iteration is: [30] train's binary_logloss: 0.529202 valid's binary_logloss: 0.659327 num_leaves, val_score: 0.647726: 55%|#####5 | 11/20 [00:08<00:08, 1.12it/s][I 2020-09-27 04:38:32,457] Trial 17 finished with value: 0.6593265762527931 and parameters: {'num_leaves': 171}. Best is trial 9 with value: 0.6477259914776615. num_leaves, val_score: 0.647726: 55%|#####5 | 11/20 [00:08<00:08, 1.12it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.003451 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.516771 valid's binary_logloss: 0.654055 Early stopping, best iteration is: [73] train's binary_logloss: 0.545672 valid's binary_logloss: 0.65329 num_leaves, val_score: 0.647726: 60%|###### | 12/20 [00:09<00:06, 1.25it/s][I 2020-09-27 04:38:33,041] Trial 18 finished with value: 0.6532896945252986 and parameters: {'num_leaves': 59}. Best is trial 9 with value: 0.6477259914776615. num_leaves, val_score: 0.647726: 60%|###### | 12/20 [00:09<00:06, 1.25it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000421 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.642488 valid's binary_logloss: 0.652362 [200] train's binary_logloss: 0.628687 valid's binary_logloss: 0.649023 [300] train's binary_logloss: 0.617877 valid's binary_logloss: 0.649151 Early stopping, best iteration is: [279] train's binary_logloss: 0.620188 valid's binary_logloss: 0.648624 num_leaves, val_score: 0.647726: 65%|######5 | 13/20 [00:10<00:05, 1.31it/s][I 2020-09-27 04:38:33,727] Trial 19 finished with value: 0.6486244754568631 and parameters: {'num_leaves': 6}. Best is trial 9 with value: 0.6477259914776615. num_leaves, val_score: 0.647726: 65%|######5 | 13/20 [00:10<00:05, 1.31it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009248 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.423908 valid's binary_logloss: 0.659234 Early stopping, best iteration is: [44] train's binary_logloss: 0.527404 valid's binary_logloss: 0.655427 num_leaves, val_score: 0.647726: 70%|####### | 14/20 [00:11<00:05, 1.07it/s][I 2020-09-27 04:38:35,066] Trial 20 finished with value: 0.6554267059223545 and parameters: {'num_leaves': 116}. Best is trial 9 with value: 0.6477259914776615. num_leaves, val_score: 0.647726: 70%|####### | 14/20 [00:11<00:05, 1.07it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004318 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.669306 valid's binary_logloss: 0.670416 [200] train's binary_logloss: 0.659732 valid's binary_logloss: 0.660321 [300] train's binary_logloss: 0.654518 valid's binary_logloss: 0.655319 [400] train's binary_logloss: 0.651377 valid's binary_logloss: 0.652506 [500] train's binary_logloss: 0.649375 valid's binary_logloss: 0.650671 [600] train's binary_logloss: 0.648029 valid's binary_logloss: 0.649863 [700] train's binary_logloss: 0.647077 valid's binary_logloss: 0.649248 [800] train's binary_logloss: 0.64636 valid's binary_logloss: 0.649015 [900] train's binary_logloss: 0.645776 valid's binary_logloss: 0.649025 [1000] train's binary_logloss: 0.645301 valid's binary_logloss: 0.649111 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.645301 valid's binary_logloss: 0.649111 num_leaves, val_score: 0.647726: 75%|#######5 | 15/20 [00:12<00:05, 1.02s/it][I 2020-09-27 04:38:36,284] Trial 21 finished with value: 0.649110563220017 and parameters: {'num_leaves': 2}. Best is trial 9 with value: 0.6477259914776615. num_leaves, val_score: 0.647726: 75%|#######5 | 15/20 [00:12<00:05, 1.02s/it][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011497 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.560366 valid's binary_logloss: 0.653399 [200] train's binary_logloss: 0.496525 valid's binary_logloss: 0.653927 Early stopping, best iteration is: [123] train's binary_logloss: 0.544369 valid's binary_logloss: 0.65141 num_leaves, val_score: 0.647726: 80%|######## | 16/20 [00:13<00:03, 1.10it/s][I 2020-09-27 04:38:36,924] Trial 22 finished with value: 0.6514101116698465 and parameters: {'num_leaves': 38}. Best is trial 9 with value: 0.6477259914776615. num_leaves, val_score: 0.647726: 80%|######## | 16/20 [00:13<00:03, 1.10it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005527 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.669306 valid's binary_logloss: 0.670416 [200] train's binary_logloss: 0.659732 valid's binary_logloss: 0.660321 [300] train's binary_logloss: 0.654518 valid's binary_logloss: 0.655319 [400] train's binary_logloss: 0.651377 valid's binary_logloss: 0.652506 [500] train's binary_logloss: 0.649375 valid's binary_logloss: 0.650671 [600] train's binary_logloss: 0.648029 valid's binary_logloss: 0.649863 [700] train's binary_logloss: 0.647077 valid's binary_logloss: 0.649248 [800] train's binary_logloss: 0.64636 valid's binary_logloss: 0.649015 [900] train's binary_logloss: 0.645776 valid's binary_logloss: 0.649025 [1000] train's binary_logloss: 0.645301 valid's binary_logloss: 0.649111 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.645301 valid's binary_logloss: 0.649111 num_leaves, val_score: 0.647726: 85%|########5 | 17/20 [00:14<00:03, 1.12s/it][I 2020-09-27 04:38:38,549] Trial 23 finished with value: 0.6491105632200171 and parameters: {'num_leaves': 2}. Best is trial 9 with value: 0.6477259914776615. num_leaves, val_score: 0.647726: 85%|########5 | 17/20 [00:14<00:03, 1.12s/it][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005133 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.53935 valid's binary_logloss: 0.652897 Early stopping, best iteration is: [92] train's binary_logloss: 0.546487 valid's binary_logloss: 0.652502 num_leaves, val_score: 0.647726: 90%|######### | 18/20 [00:15<00:01, 1.04it/s][I 2020-09-27 04:38:39,143] Trial 24 finished with value: 0.6525017174749979 and parameters: {'num_leaves': 48}. Best is trial 9 with value: 0.6477259914776615. num_leaves, val_score: 0.647726: 90%|######### | 18/20 [00:15<00:01, 1.04it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.003834 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.669306 valid's binary_logloss: 0.670416 [200] train's binary_logloss: 0.659732 valid's binary_logloss: 0.660321 [300] train's binary_logloss: 0.654518 valid's binary_logloss: 0.655319 [400] train's binary_logloss: 0.651377 valid's binary_logloss: 0.652506 [500] train's binary_logloss: 0.649375 valid's binary_logloss: 0.650671 [600] train's binary_logloss: 0.648029 valid's binary_logloss: 0.649863 [700] train's binary_logloss: 0.647077 valid's binary_logloss: 0.649248 [800] train's binary_logloss: 0.64636 valid's binary_logloss: 0.649015 [900] train's binary_logloss: 0.645776 valid's binary_logloss: 0.649025 [1000] train's binary_logloss: 0.645301 valid's binary_logloss: 0.649111 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.645301 valid's binary_logloss: 0.649111 num_leaves, val_score: 0.647726: 95%|#########5| 19/20 [00:16<00:01, 1.02s/it][I 2020-09-27 04:38:40,291] Trial 25 finished with value: 0.649110563220017 and parameters: {'num_leaves': 2}. Best is trial 9 with value: 0.6477259914776615. num_leaves, val_score: 0.647726: 95%|#########5| 19/20 [00:16<00:01, 1.02s/it][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000543 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.468108 valid's binary_logloss: 0.654681 Early stopping, best iteration is: [69] train's binary_logloss: 0.512521 valid's binary_logloss: 0.652052 num_leaves, val_score: 0.647726: 100%|##########| 20/20 [00:17<00:00, 1.05it/s][I 2020-09-27 04:38:41,075] Trial 26 finished with value: 0.652051918071773 and parameters: {'num_leaves': 86}. Best is trial 9 with value: 0.6477259914776615. num_leaves, val_score: 0.647726: 100%|##########| 20/20 [00:17<00:00, 1.15it/s] bagging, val_score: 0.647726: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.003535 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.63651 valid's binary_logloss: 0.650977 [200] train's binary_logloss: 0.620185 valid's binary_logloss: 0.652728 Early stopping, best iteration is: [150] train's binary_logloss: 0.627724 valid's binary_logloss: 0.650131 bagging, val_score: 0.647726: 10%|# | 1/10 [00:00<00:04, 1.99it/s][I 2020-09-27 04:38:41,600] Trial 27 finished with value: 0.650131071613353 and parameters: {'bagging_fraction': 0.4586199697400575, 'bagging_freq': 3}. Best is trial 27 with value: 0.650131071613353. bagging, val_score: 0.647726: 10%|# | 1/10 [00:00<00:04, 1.99it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000853 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.635382 valid's binary_logloss: 0.651145 [200] train's binary_logloss: 0.61788 valid's binary_logloss: 0.647841 [300] train's binary_logloss: 0.603222 valid's binary_logloss: 0.647911 [400] train's binary_logloss: 0.589945 valid's binary_logloss: 0.647612 Early stopping, best iteration is: [340] train's binary_logloss: 0.597917 valid's binary_logloss: 0.647006 bagging, val_score: 0.647006: 20%|## | 2/10 [00:01<00:05, 1.33it/s][I 2020-09-27 04:38:42,919] Trial 28 finished with value: 0.6470063208679462 and parameters: {'bagging_fraction': 0.9958662267298061, 'bagging_freq': 7}. Best is trial 28 with value: 0.6470063208679462. bagging, val_score: 0.647006: 20%|## | 2/10 [00:01<00:05, 1.33it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.003784 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.635259 valid's binary_logloss: 0.651142 [200] train's binary_logloss: 0.617565 valid's binary_logloss: 0.64962 Early stopping, best iteration is: [195] train's binary_logloss: 0.618361 valid's binary_logloss: 0.649343 bagging, val_score: 0.647006: 30%|### | 3/10 [00:02<00:04, 1.45it/s][I 2020-09-27 04:38:43,462] Trial 29 finished with value: 0.6493431744945917 and parameters: {'bagging_fraction': 0.9799916779746923, 'bagging_freq': 7}. Best is trial 28 with value: 0.6470063208679462. bagging, val_score: 0.647006: 30%|### | 3/10 [00:02<00:04, 1.45it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000779 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.635241 valid's binary_logloss: 0.651269 [200] train's binary_logloss: 0.618036 valid's binary_logloss: 0.649672 [300] train's binary_logloss: 0.603932 valid's binary_logloss: 0.649408 [400] train's binary_logloss: 0.590656 valid's binary_logloss: 0.648685 Early stopping, best iteration is: [368] train's binary_logloss: 0.594891 valid's binary_logloss: 0.648347 bagging, val_score: 0.647006: 40%|#### | 4/10 [00:03<00:04, 1.37it/s][I 2020-09-27 04:38:44,294] Trial 30 finished with value: 0.6483471798474496 and parameters: {'bagging_fraction': 0.9986614103818544, 'bagging_freq': 7}. Best is trial 28 with value: 0.6470063208679462. bagging, val_score: 0.647006: 40%|#### | 4/10 [00:03<00:04, 1.37it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007385 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.63573 valid's binary_logloss: 0.649902 [200] train's binary_logloss: 0.618646 valid's binary_logloss: 0.650492 Early stopping, best iteration is: [170] train's binary_logloss: 0.623387 valid's binary_logloss: 0.648591 bagging, val_score: 0.647006: 50%|##### | 5/10 [00:03<00:03, 1.52it/s][I 2020-09-27 04:38:44,781] Trial 31 finished with value: 0.6485905878341305 and parameters: {'bagging_fraction': 0.7328660882461872, 'bagging_freq': 5}. Best is trial 28 with value: 0.6470063208679462. bagging, val_score: 0.647006: 50%|##### | 5/10 [00:03<00:03, 1.52it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000825 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.635272 valid's binary_logloss: 0.650422 [200] train's binary_logloss: 0.61747 valid's binary_logloss: 0.648941 [300] train's binary_logloss: 0.602462 valid's binary_logloss: 0.650582 Early stopping, best iteration is: [227] train's binary_logloss: 0.613139 valid's binary_logloss: 0.648451 bagging, val_score: 0.647006: 60%|###### | 6/10 [00:04<00:02, 1.49it/s][I 2020-09-27 04:38:45,485] Trial 32 finished with value: 0.6484510872456424 and parameters: {'bagging_fraction': 0.7900181913598195, 'bagging_freq': 1}. Best is trial 28 with value: 0.6470063208679462. bagging, val_score: 0.647006: 60%|###### | 6/10 [00:04<00:02, 1.49it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.003362 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.636504 valid's binary_logloss: 0.650927 [200] train's binary_logloss: 0.620547 valid's binary_logloss: 0.651909 Early stopping, best iteration is: [115] train's binary_logloss: 0.633633 valid's binary_logloss: 0.6507 bagging, val_score: 0.647006: 70%|####### | 7/10 [00:05<00:02, 1.40it/s][I 2020-09-27 04:38:46,298] Trial 33 finished with value: 0.6506995210705391 and parameters: {'bagging_fraction': 0.47267376312879716, 'bagging_freq': 5}. Best is trial 28 with value: 0.6470063208679462. bagging, val_score: 0.647006: 70%|####### | 7/10 [00:05<00:02, 1.40it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005528 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.635448 valid's binary_logloss: 0.651834 [200] train's binary_logloss: 0.617764 valid's binary_logloss: 0.650603 [300] train's binary_logloss: 0.603475 valid's binary_logloss: 0.652326 Early stopping, best iteration is: [226] train's binary_logloss: 0.614069 valid's binary_logloss: 0.649886 bagging, val_score: 0.647006: 80%|######## | 8/10 [00:05<00:01, 1.47it/s][I 2020-09-27 04:38:46,894] Trial 34 finished with value: 0.649885537943031 and parameters: {'bagging_fraction': 0.876875549392964, 'bagging_freq': 7}. Best is trial 28 with value: 0.6470063208679462. bagging, val_score: 0.647006: 80%|######## | 8/10 [00:05<00:01, 1.47it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000544 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.635851 valid's binary_logloss: 0.649962 [200] train's binary_logloss: 0.6186 valid's binary_logloss: 0.647048 [300] train's binary_logloss: 0.603846 valid's binary_logloss: 0.649943 Early stopping, best iteration is: [207] train's binary_logloss: 0.617528 valid's binary_logloss: 0.646928 bagging, val_score: 0.646928: 90%|######### | 9/10 [00:06<00:00, 1.56it/s][I 2020-09-27 04:38:47,444] Trial 35 finished with value: 0.6469279610503994 and parameters: {'bagging_fraction': 0.6178680364111143, 'bagging_freq': 3}. Best is trial 35 with value: 0.6469279610503994. bagging, val_score: 0.646928: 90%|######### | 9/10 [00:06<00:00, 1.56it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000490 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.635909 valid's binary_logloss: 0.648293 [200] train's binary_logloss: 0.618922 valid's binary_logloss: 0.649973 Early stopping, best iteration is: [149] train's binary_logloss: 0.627266 valid's binary_logloss: 0.647531 bagging, val_score: 0.646928: 100%|##########| 10/10 [00:06<00:00, 1.68it/s][I 2020-09-27 04:38:47,929] Trial 36 finished with value: 0.6475307026936266 and parameters: {'bagging_fraction': 0.5956628615183736, 'bagging_freq': 2}. Best is trial 35 with value: 0.6469279610503994. bagging, val_score: 0.646928: 100%|##########| 10/10 [00:06<00:00, 1.46it/s] feature_fraction_stage2, val_score: 0.646928: 0%| | 0/6 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000428 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.634952 valid's binary_logloss: 0.65056 [200] train's binary_logloss: 0.617682 valid's binary_logloss: 0.649553 [300] train's binary_logloss: 0.603161 valid's binary_logloss: 0.652262 Early stopping, best iteration is: [241] train's binary_logloss: 0.611549 valid's binary_logloss: 0.649321 feature_fraction_stage2, val_score: 0.646928: 17%|#6 | 1/6 [00:00<00:03, 1.60it/s][I 2020-09-27 04:38:48,575] Trial 37 finished with value: 0.6493214800336689 and parameters: {'feature_fraction': 0.748}. Best is trial 37 with value: 0.6493214800336689. feature_fraction_stage2, val_score: 0.646928: 17%|#6 | 1/6 [00:00<00:03, 1.60it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001333 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.635533 valid's binary_logloss: 0.650663 [200] train's binary_logloss: 0.618659 valid's binary_logloss: 0.651432 Early stopping, best iteration is: [141] train's binary_logloss: 0.628172 valid's binary_logloss: 0.650017 feature_fraction_stage2, val_score: 0.646928: 33%|###3 | 2/6 [00:01<00:02, 1.67it/s][I 2020-09-27 04:38:49,111] Trial 38 finished with value: 0.6500171564584806 and parameters: {'feature_fraction': 0.652}. Best is trial 37 with value: 0.6493214800336689. feature_fraction_stage2, val_score: 0.646928: 33%|###3 | 2/6 [00:01<00:02, 1.67it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000674 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.635851 valid's binary_logloss: 0.649962 [200] train's binary_logloss: 0.6186 valid's binary_logloss: 0.647048 [300] train's binary_logloss: 0.603846 valid's binary_logloss: 0.649943 Early stopping, best iteration is: [207] train's binary_logloss: 0.617528 valid's binary_logloss: 0.646928 feature_fraction_stage2, val_score: 0.646928: 50%|##### | 3/6 [00:02<00:02, 1.33it/s][I 2020-09-27 04:38:50,213] Trial 39 finished with value: 0.6469279610503994 and parameters: {'feature_fraction': 0.6839999999999999}. Best is trial 39 with value: 0.6469279610503994. feature_fraction_stage2, val_score: 0.646928: 50%|##### | 3/6 [00:02<00:02, 1.33it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000915 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.635444 valid's binary_logloss: 0.649405 [200] train's binary_logloss: 0.618068 valid's binary_logloss: 0.648194 Early stopping, best iteration is: [173] train's binary_logloss: 0.622667 valid's binary_logloss: 0.647565 feature_fraction_stage2, val_score: 0.646928: 67%|######6 | 4/6 [00:02<00:01, 1.46it/s][I 2020-09-27 04:38:50,749] Trial 40 finished with value: 0.6475654368255805 and parameters: {'feature_fraction': 0.7799999999999999}. Best is trial 39 with value: 0.6469279610503994. feature_fraction_stage2, val_score: 0.646928: 67%|######6 | 4/6 [00:02<00:01, 1.46it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004998 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.63624 valid's binary_logloss: 0.650875 [200] train's binary_logloss: 0.619436 valid's binary_logloss: 0.649352 Early stopping, best iteration is: [165] train's binary_logloss: 0.62485 valid's binary_logloss: 0.648991 feature_fraction_stage2, val_score: 0.646928: 83%|########3 | 5/6 [00:03<00:00, 1.59it/s][I 2020-09-27 04:38:51,248] Trial 41 finished with value: 0.6489914374612618 and parameters: {'feature_fraction': 0.62}. Best is trial 39 with value: 0.6469279610503994. feature_fraction_stage2, val_score: 0.646928: 83%|########3 | 5/6 [00:03<00:00, 1.59it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000866 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.634952 valid's binary_logloss: 0.65056 [200] train's binary_logloss: 0.617682 valid's binary_logloss: 0.649553 [300] train's binary_logloss: 0.603161 valid's binary_logloss: 0.652262 Early stopping, best iteration is: [241] train's binary_logloss: 0.611549 valid's binary_logloss: 0.649321 feature_fraction_stage2, val_score: 0.646928: 100%|##########| 6/6 [00:03<00:00, 1.56it/s][I 2020-09-27 04:38:51,915] Trial 42 finished with value: 0.6493214800336689 and parameters: {'feature_fraction': 0.716}. Best is trial 39 with value: 0.6469279610503994. feature_fraction_stage2, val_score: 0.646928: 100%|##########| 6/6 [00:03<00:00, 1.51it/s] regularization_factors, val_score: 0.646928: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000644 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.635945 valid's binary_logloss: 0.65002 [200] train's binary_logloss: 0.61936 valid's binary_logloss: 0.648569 Early stopping, best iteration is: [161] train's binary_logloss: 0.625243 valid's binary_logloss: 0.647758 regularization_factors, val_score: 0.646928: 5%|5 | 1/20 [00:00<00:10, 1.73it/s][I 2020-09-27 04:38:52,510] Trial 43 finished with value: 0.6477575542610232 and parameters: {'lambda_l1': 0.5305532619182826, 'lambda_l2': 2.2940234848178292e-05}. Best is trial 43 with value: 0.6477575542610232. regularization_factors, val_score: 0.646928: 5%|5 | 1/20 [00:00<00:10, 1.73it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004925 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.637159 valid's binary_logloss: 0.648561 [200] train's binary_logloss: 0.62222 valid's binary_logloss: 0.646184 [300] train's binary_logloss: 0.610053 valid's binary_logloss: 0.646865 Early stopping, best iteration is: [250] train's binary_logloss: 0.616221 valid's binary_logloss: 0.645506 regularization_factors, val_score: 0.645506: 10%|# | 2/20 [00:01<00:12, 1.48it/s][I 2020-09-27 04:38:53,419] Trial 44 finished with value: 0.6455057317841788 and parameters: {'lambda_l1': 1.5729907152074253e-08, 'lambda_l2': 6.408315304180313}. Best is trial 44 with value: 0.6455057317841788. regularization_factors, val_score: 0.645506: 10%|# | 2/20 [00:01<00:12, 1.48it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010288 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.637347 valid's binary_logloss: 0.64928 [200] train's binary_logloss: 0.623416 valid's binary_logloss: 0.647003 [300] train's binary_logloss: 0.611546 valid's binary_logloss: 0.648015 Early stopping, best iteration is: [225] train's binary_logloss: 0.620614 valid's binary_logloss: 0.646489 regularization_factors, val_score: 0.645506: 15%|#5 | 3/20 [00:02<00:12, 1.34it/s][I 2020-09-27 04:38:54,330] Trial 45 finished with value: 0.6464889390748203 and parameters: {'lambda_l1': 1.792278679780671e-08, 'lambda_l2': 8.810233526119564}. Best is trial 44 with value: 0.6455057317841788. regularization_factors, val_score: 0.645506: 15%|#5 | 3/20 [00:02<00:12, 1.34it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000443 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.63666 valid's binary_logloss: 0.649464 [200] train's binary_logloss: 0.62174 valid's binary_logloss: 0.647351 [300] train's binary_logloss: 0.609288 valid's binary_logloss: 0.648443 Early stopping, best iteration is: [254] train's binary_logloss: 0.615011 valid's binary_logloss: 0.646435 regularization_factors, val_score: 0.645506: 20%|## | 4/20 [00:03<00:11, 1.37it/s][I 2020-09-27 04:38:55,027] Trial 46 finished with value: 0.6464351970725228 and parameters: {'lambda_l1': 2.2856982982765707e-08, 'lambda_l2': 4.5894526330871095}. Best is trial 44 with value: 0.6455057317841788. regularization_factors, val_score: 0.645506: 20%|## | 4/20 [00:03<00:11, 1.37it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.007799 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.637474 valid's binary_logloss: 0.648583 [200] train's binary_logloss: 0.62337 valid's binary_logloss: 0.645859 [300] train's binary_logloss: 0.61123 valid's binary_logloss: 0.645293 [400] train's binary_logloss: 0.599843 valid's binary_logloss: 0.646225 Early stopping, best iteration is: [317] train's binary_logloss: 0.609063 valid's binary_logloss: 0.644762 regularization_factors, val_score: 0.644762: 25%|##5 | 5/20 [00:03<00:11, 1.32it/s][I 2020-09-27 04:38:55,839] Trial 47 finished with value: 0.6447624933131433 and parameters: {'lambda_l1': 1.2755111716437187e-08, 'lambda_l2': 9.850714874340618}. Best is trial 47 with value: 0.6447624933131433. regularization_factors, val_score: 0.644762: 25%|##5 | 5/20 [00:03<00:11, 1.32it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004054 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.63754 valid's binary_logloss: 0.648978 [200] train's binary_logloss: 0.623228 valid's binary_logloss: 0.646323 [300] train's binary_logloss: 0.611727 valid's binary_logloss: 0.64615 [400] train's binary_logloss: 0.600473 valid's binary_logloss: 0.646062 Early stopping, best iteration is: [347] train's binary_logloss: 0.606258 valid's binary_logloss: 0.645207 regularization_factors, val_score: 0.644762: 30%|### | 6/20 [00:04<00:11, 1.26it/s][I 2020-09-27 04:38:56,723] Trial 48 finished with value: 0.6452071554754294 and parameters: {'lambda_l1': 1.3011629397041354e-08, 'lambda_l2': 9.988192496777456}. Best is trial 47 with value: 0.6447624933131433. regularization_factors, val_score: 0.644762: 30%|### | 6/20 [00:04<00:11, 1.26it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007536 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.6371 valid's binary_logloss: 0.6496 [200] train's binary_logloss: 0.622819 valid's binary_logloss: 0.647511 [300] train's binary_logloss: 0.610769 valid's binary_logloss: 0.647971 Early stopping, best iteration is: [232] train's binary_logloss: 0.618821 valid's binary_logloss: 0.646619 regularization_factors, val_score: 0.644762: 35%|###5 | 7/20 [00:05<00:11, 1.14it/s][I 2020-09-27 04:38:57,789] Trial 49 finished with value: 0.6466188865283253 and parameters: {'lambda_l1': 1.5462932963049097e-08, 'lambda_l2': 8.71269927175754}. Best is trial 47 with value: 0.6447624933131433. regularization_factors, val_score: 0.644762: 35%|###5 | 7/20 [00:05<00:11, 1.14it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004606 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.637133 valid's binary_logloss: 0.649344 [200] train's binary_logloss: 0.622572 valid's binary_logloss: 0.647963 [300] train's binary_logloss: 0.61055 valid's binary_logloss: 0.647491 Early stopping, best iteration is: [255] train's binary_logloss: 0.616042 valid's binary_logloss: 0.646564 regularization_factors, val_score: 0.644762: 40%|#### | 8/20 [00:06<00:09, 1.21it/s][I 2020-09-27 04:38:58,493] Trial 50 finished with value: 0.6465642008648574 and parameters: {'lambda_l1': 1.1089576217805702e-08, 'lambda_l2': 7.312599056585437}. Best is trial 47 with value: 0.6447624933131433. regularization_factors, val_score: 0.644762: 40%|#### | 8/20 [00:06<00:09, 1.21it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000434 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.637479 valid's binary_logloss: 0.648584 [200] train's binary_logloss: 0.62341 valid's binary_logloss: 0.645728 [300] train's binary_logloss: 0.611552 valid's binary_logloss: 0.64603 Early stopping, best iteration is: [254] train's binary_logloss: 0.617107 valid's binary_logloss: 0.644789 regularization_factors, val_score: 0.644762: 45%|####5 | 9/20 [00:07<00:08, 1.27it/s][I 2020-09-27 04:38:59,196] Trial 51 finished with value: 0.6447887069638869 and parameters: {'lambda_l1': 1.7016389032742813e-08, 'lambda_l2': 9.898211761649065}. Best is trial 47 with value: 0.6447624933131433. regularization_factors, val_score: 0.644762: 45%|####5 | 9/20 [00:07<00:08, 1.27it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007899 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.637278 valid's binary_logloss: 0.649089 [200] train's binary_logloss: 0.622576 valid's binary_logloss: 0.647208 [300] train's binary_logloss: 0.610418 valid's binary_logloss: 0.648157 Early stopping, best iteration is: [255] train's binary_logloss: 0.615802 valid's binary_logloss: 0.646398 regularization_factors, val_score: 0.644762: 50%|##### | 10/20 [00:07<00:07, 1.31it/s][I 2020-09-27 04:38:59,894] Trial 52 finished with value: 0.6463976805891822 and parameters: {'lambda_l1': 2.2240739211030176e-08, 'lambda_l2': 7.404538655662823}. Best is trial 47 with value: 0.6447624933131433. regularization_factors, val_score: 0.644762: 50%|##### | 10/20 [00:07<00:07, 1.31it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000572 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.637182 valid's binary_logloss: 0.647682 [200] train's binary_logloss: 0.623009 valid's binary_logloss: 0.646084 [300] train's binary_logloss: 0.611035 valid's binary_logloss: 0.645152 Early stopping, best iteration is: [260] train's binary_logloss: 0.615822 valid's binary_logloss: 0.643798 regularization_factors, val_score: 0.643798: 55%|#####5 | 11/20 [00:08<00:06, 1.35it/s][I 2020-09-27 04:39:00,581] Trial 53 finished with value: 0.6437977363337001 and parameters: {'lambda_l1': 2.039143182350295e-08, 'lambda_l2': 9.400199053113356}. Best is trial 53 with value: 0.6437977363337001. regularization_factors, val_score: 0.643798: 55%|#####5 | 11/20 [00:08<00:06, 1.35it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004032 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.637206 valid's binary_logloss: 0.647684 [200] train's binary_logloss: 0.623151 valid's binary_logloss: 0.646089 [300] train's binary_logloss: 0.611403 valid's binary_logloss: 0.645967 Early stopping, best iteration is: [255] train's binary_logloss: 0.616582 valid's binary_logloss: 0.644531 regularization_factors, val_score: 0.643798: 60%|###### | 12/20 [00:09<00:06, 1.17it/s][I 2020-09-27 04:39:01,697] Trial 54 finished with value: 0.6445311936163728 and parameters: {'lambda_l1': 1.6270643281831753e-08, 'lambda_l2': 9.628307902916124}. Best is trial 53 with value: 0.6437977363337001. regularization_factors, val_score: 0.643798: 60%|###### | 12/20 [00:09<00:06, 1.17it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000874 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.637281 valid's binary_logloss: 0.649044 [200] train's binary_logloss: 0.622584 valid's binary_logloss: 0.647021 [300] train's binary_logloss: 0.610763 valid's binary_logloss: 0.647525 Early stopping, best iteration is: [254] train's binary_logloss: 0.616357 valid's binary_logloss: 0.645998 regularization_factors, val_score: 0.643798: 65%|######5 | 13/20 [00:10<00:05, 1.22it/s][I 2020-09-27 04:39:02,449] Trial 55 finished with value: 0.6459984456397755 and parameters: {'lambda_l1': 1.0240697224966498e-08, 'lambda_l2': 8.012028500403863}. Best is trial 53 with value: 0.6437977363337001. regularization_factors, val_score: 0.643798: 65%|######5 | 13/20 [00:10<00:05, 1.22it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000558 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.635858 valid's binary_logloss: 0.649961 [200] train's binary_logloss: 0.618625 valid's binary_logloss: 0.647047 [300] train's binary_logloss: 0.603932 valid's binary_logloss: 0.648748 Early stopping, best iteration is: [207] train's binary_logloss: 0.617554 valid's binary_logloss: 0.646927 regularization_factors, val_score: 0.643798: 70%|####### | 14/20 [00:11<00:04, 1.32it/s][I 2020-09-27 04:39:03,059] Trial 56 finished with value: 0.6469267717887545 and parameters: {'lambda_l1': 1.2959832015054276e-06, 'lambda_l2': 0.01861979600536283}. Best is trial 53 with value: 0.6437977363337001. regularization_factors, val_score: 0.643798: 70%|####### | 14/20 [00:11<00:04, 1.32it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010899 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.635708 valid's binary_logloss: 0.649563 [200] train's binary_logloss: 0.61871 valid's binary_logloss: 0.647652 [300] train's binary_logloss: 0.604402 valid's binary_logloss: 0.650241 Early stopping, best iteration is: [206] train's binary_logloss: 0.6179 valid's binary_logloss: 0.647417 regularization_factors, val_score: 0.643798: 75%|#######5 | 15/20 [00:11<00:03, 1.43it/s][I 2020-09-27 04:39:03,628] Trial 57 finished with value: 0.6474166124124395 and parameters: {'lambda_l1': 5.765532893767189e-07, 'lambda_l2': 0.14788147344128014}. Best is trial 53 with value: 0.6437977363337001. regularization_factors, val_score: 0.643798: 75%|#######5 | 15/20 [00:11<00:03, 1.43it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005282 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.635911 valid's binary_logloss: 0.650092 [200] train's binary_logloss: 0.618697 valid's binary_logloss: 0.649148 [300] train's binary_logloss: 0.604075 valid's binary_logloss: 0.650674 Early stopping, best iteration is: [250] train's binary_logloss: 0.611662 valid's binary_logloss: 0.648499 regularization_factors, val_score: 0.643798: 80%|######## | 16/20 [00:12<00:02, 1.44it/s][I 2020-09-27 04:39:04,302] Trial 58 finished with value: 0.6484991507925789 and parameters: {'lambda_l1': 5.203220654613759e-07, 'lambda_l2': 0.30810693370096226}. Best is trial 53 with value: 0.6437977363337001. regularization_factors, val_score: 0.643798: 80%|######## | 16/20 [00:12<00:02, 1.44it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004372 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.63618 valid's binary_logloss: 0.650527 [200] train's binary_logloss: 0.619801 valid's binary_logloss: 0.649123 Early stopping, best iteration is: [160] train's binary_logloss: 0.625979 valid's binary_logloss: 0.648367 regularization_factors, val_score: 0.643798: 85%|########5 | 17/20 [00:13<00:02, 1.37it/s][I 2020-09-27 04:39:05,120] Trial 59 finished with value: 0.6483666505535219 and parameters: {'lambda_l1': 1.1681262794154124e-07, 'lambda_l2': 0.9991529121249367}. Best is trial 53 with value: 0.6437977363337001. regularization_factors, val_score: 0.643798: 85%|########5 | 17/20 [00:13<00:02, 1.37it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008489 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.635851 valid's binary_logloss: 0.649962 [200] train's binary_logloss: 0.6186 valid's binary_logloss: 0.647048 [300] train's binary_logloss: 0.603848 valid's binary_logloss: 0.649943 Early stopping, best iteration is: [207] train's binary_logloss: 0.617528 valid's binary_logloss: 0.646928 regularization_factors, val_score: 0.643798: 90%|######### | 18/20 [00:13<00:01, 1.40it/s][I 2020-09-27 04:39:05,803] Trial 60 finished with value: 0.6469279610385475 and parameters: {'lambda_l1': 1.213400855943739e-08, 'lambda_l2': 1.534183918800275e-07}. Best is trial 53 with value: 0.6437977363337001. regularization_factors, val_score: 0.643798: 90%|######### | 18/20 [00:13<00:01, 1.40it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009995 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.637219 valid's binary_logloss: 0.650017 [200] train's binary_logloss: 0.622187 valid's binary_logloss: 0.647211 [300] train's binary_logloss: 0.609711 valid's binary_logloss: 0.647263 Early stopping, best iteration is: [254] train's binary_logloss: 0.615481 valid's binary_logloss: 0.64613 regularization_factors, val_score: 0.643798: 95%|#########5| 19/20 [00:14<00:00, 1.44it/s][I 2020-09-27 04:39:06,446] Trial 61 finished with value: 0.6461303865450015 and parameters: {'lambda_l1': 1.0232750932997174e-08, 'lambda_l2': 5.61104076150738}. Best is trial 53 with value: 0.6437977363337001. regularization_factors, val_score: 0.643798: 95%|#########5| 19/20 [00:14<00:00, 1.44it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000915 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.637448 valid's binary_logloss: 0.649065 [200] train's binary_logloss: 0.623392 valid's binary_logloss: 0.647489 [300] train's binary_logloss: 0.611817 valid's binary_logloss: 0.647888 Early stopping, best iteration is: [236] train's binary_logloss: 0.61913 valid's binary_logloss: 0.647044 regularization_factors, val_score: 0.643798: 100%|##########| 20/20 [00:15<00:00, 1.48it/s][I 2020-09-27 04:39:07,086] Trial 62 finished with value: 0.6470438910627877 and parameters: {'lambda_l1': 1.4832304242797658e-08, 'lambda_l2': 9.997088293767975}. Best is trial 53 with value: 0.6437977363337001. regularization_factors, val_score: 0.643798: 100%|##########| 20/20 [00:15<00:00, 1.32it/s] min_data_in_leaf, val_score: 0.643798: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.003521 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.637289 valid's binary_logloss: 0.648183 [200] train's binary_logloss: 0.62309 valid's binary_logloss: 0.64706 [300] train's binary_logloss: 0.611616 valid's binary_logloss: 0.64692 Early stopping, best iteration is: [254] train's binary_logloss: 0.616891 valid's binary_logloss: 0.645943 min_data_in_leaf, val_score: 0.643798: 20%|## | 1/5 [00:00<00:02, 1.35it/s][I 2020-09-27 04:39:07,843] Trial 63 finished with value: 0.6459430156769583 and parameters: {'min_child_samples': 50}. Best is trial 63 with value: 0.6459430156769583. min_data_in_leaf, val_score: 0.643798: 20%|## | 1/5 [00:00<00:02, 1.35it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000832 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.637182 valid's binary_logloss: 0.647682 [200] train's binary_logloss: 0.622869 valid's binary_logloss: 0.646623 [300] train's binary_logloss: 0.611124 valid's binary_logloss: 0.646202 Early stopping, best iteration is: [254] train's binary_logloss: 0.616639 valid's binary_logloss: 0.645162 min_data_in_leaf, val_score: 0.643798: 40%|#### | 2/5 [00:01<00:02, 1.27it/s][I 2020-09-27 04:39:08,731] Trial 64 finished with value: 0.6451618189346408 and parameters: {'min_child_samples': 10}. Best is trial 64 with value: 0.6451618189346408. min_data_in_leaf, val_score: 0.643798: 40%|#### | 2/5 [00:01<00:02, 1.27it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012841 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.637182 valid's binary_logloss: 0.647682 [200] train's binary_logloss: 0.623074 valid's binary_logloss: 0.64581 [300] train's binary_logloss: 0.61124 valid's binary_logloss: 0.645222 Early stopping, best iteration is: [255] train's binary_logloss: 0.616649 valid's binary_logloss: 0.643974 min_data_in_leaf, val_score: 0.643798: 60%|###### | 3/5 [00:02<00:01, 1.11it/s][I 2020-09-27 04:39:09,890] Trial 65 finished with value: 0.6439740196693694 and parameters: {'min_child_samples': 25}. Best is trial 65 with value: 0.6439740196693694. min_data_in_leaf, val_score: 0.643798: 60%|###### | 3/5 [00:02<00:01, 1.11it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004742 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.637182 valid's binary_logloss: 0.647682 [200] train's binary_logloss: 0.622869 valid's binary_logloss: 0.646623 [300] train's binary_logloss: 0.610951 valid's binary_logloss: 0.646227 Early stopping, best iteration is: [243] train's binary_logloss: 0.617796 valid's binary_logloss: 0.645499 min_data_in_leaf, val_score: 0.643798: 80%|######## | 4/5 [00:03<00:00, 1.19it/s][I 2020-09-27 04:39:10,600] Trial 66 finished with value: 0.6454986987713198 and parameters: {'min_child_samples': 5}. Best is trial 65 with value: 0.6439740196693694. min_data_in_leaf, val_score: 0.643798: 80%|######## | 4/5 [00:03<00:00, 1.19it/s][LightGBM] [Info] Number of positive: 13186, number of negative: 12813 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000441 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.507173 -> initscore=0.028695 [LightGBM] [Info] Start training from score 0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.637729 valid's binary_logloss: 0.649258 [200] train's binary_logloss: 0.62325 valid's binary_logloss: 0.646308 [300] train's binary_logloss: 0.611957 valid's binary_logloss: 0.647641 Early stopping, best iteration is: [238] train's binary_logloss: 0.618817 valid's binary_logloss: 0.645531 min_data_in_leaf, val_score: 0.643798: 100%|##########| 5/5 [00:04<00:00, 1.26it/s][I 2020-09-27 04:39:11,288] Trial 67 finished with value: 0.6455312591688435 and parameters: {'min_child_samples': 100}. Best is trial 65 with value: 0.6439740196693694. min_data_in_leaf, val_score: 0.643798: 100%|##########| 5/5 [00:04<00:00, 1.19it/s] Fold : 1 [I 2020-09-27 04:39:11,344] A new study created in memory with name: no-name-8ec748bd-7c24-485d-8dfe-5099f051a9d9 feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001101 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.575663 valid's binary_logloss: 0.661357 Early stopping, best iteration is: [56] train's binary_logloss: 0.605835 valid's binary_logloss: 0.659163 feature_fraction, val_score: 0.659163: 14%|#4 | 1/7 [00:00<00:03, 1.89it/s][I 2020-09-27 04:39:11,890] Trial 0 finished with value: 0.6591626166519308 and parameters: {'feature_fraction': 0.7}. Best is trial 0 with value: 0.6591626166519308. feature_fraction, val_score: 0.659163: 14%|#4 | 1/7 [00:00<00:03, 1.89it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000533 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.577716 valid's binary_logloss: 0.658007 Early stopping, best iteration is: [72] train's binary_logloss: 0.596211 valid's binary_logloss: 0.65673 feature_fraction, val_score: 0.656730: 29%|##8 | 2/7 [00:00<00:02, 1.98it/s][I 2020-09-27 04:39:12,340] Trial 1 finished with value: 0.6567303283064773 and parameters: {'feature_fraction': 0.5}. Best is trial 1 with value: 0.6567303283064773. feature_fraction, val_score: 0.656730: 29%|##8 | 2/7 [00:00<00:02, 1.98it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001310 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.576271 valid's binary_logloss: 0.660374 Early stopping, best iteration is: [66] train's binary_logloss: 0.598062 valid's binary_logloss: 0.658574 feature_fraction, val_score: 0.656730: 43%|####2 | 3/7 [00:01<00:02, 1.53it/s][I 2020-09-27 04:39:13,335] Trial 2 finished with value: 0.6585740014120882 and parameters: {'feature_fraction': 0.6}. Best is trial 1 with value: 0.6567303283064773. feature_fraction, val_score: 0.656730: 43%|####2 | 3/7 [00:01<00:02, 1.53it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000370 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.582075 valid's binary_logloss: 0.6565 Early stopping, best iteration is: [94] train's binary_logloss: 0.58569 valid's binary_logloss: 0.655992 feature_fraction, val_score: 0.655992: 57%|#####7 | 4/7 [00:02<00:01, 1.67it/s][I 2020-09-27 04:39:13,802] Trial 3 finished with value: 0.6559924908506093 and parameters: {'feature_fraction': 0.4}. Best is trial 3 with value: 0.6559924908506093. feature_fraction, val_score: 0.655992: 57%|#####7 | 4/7 [00:02<00:01, 1.67it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000609 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.571094 valid's binary_logloss: 0.661412 Early stopping, best iteration is: [54] train's binary_logloss: 0.604656 valid's binary_logloss: 0.657544 feature_fraction, val_score: 0.655992: 71%|#######1 | 5/7 [00:02<00:01, 1.78it/s][I 2020-09-27 04:39:14,285] Trial 4 finished with value: 0.6575438872927727 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 3 with value: 0.6559924908506093. feature_fraction, val_score: 0.655992: 71%|#######1 | 5/7 [00:02<00:01, 1.78it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000923 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.569935 valid's binary_logloss: 0.659342 Early stopping, best iteration is: [54] train's binary_logloss: 0.602845 valid's binary_logloss: 0.658926 feature_fraction, val_score: 0.655992: 86%|########5 | 6/7 [00:03<00:00, 1.86it/s][I 2020-09-27 04:39:14,768] Trial 5 finished with value: 0.658926225620644 and parameters: {'feature_fraction': 1.0}. Best is trial 3 with value: 0.6559924908506093. feature_fraction, val_score: 0.655992: 86%|########5 | 6/7 [00:03<00:00, 1.86it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000465 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.574321 valid's binary_logloss: 0.659912 Early stopping, best iteration is: [66] train's binary_logloss: 0.595979 valid's binary_logloss: 0.657826 feature_fraction, val_score: 0.655992: 100%|##########| 7/7 [00:03<00:00, 1.91it/s][I 2020-09-27 04:39:15,259] Trial 6 finished with value: 0.65782592600512 and parameters: {'feature_fraction': 0.8}. Best is trial 3 with value: 0.6559924908506093. feature_fraction, val_score: 0.655992: 100%|##########| 7/7 [00:03<00:00, 1.79it/s] num_leaves, val_score: 0.655992: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000434 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.631152 valid's binary_logloss: 0.657113 [200] train's binary_logloss: 0.611555 valid's binary_logloss: 0.657151 Early stopping, best iteration is: [114] train's binary_logloss: 0.627933 valid's binary_logloss: 0.656291 num_leaves, val_score: 0.655992: 5%|5 | 1/20 [00:00<00:07, 2.47it/s][I 2020-09-27 04:39:15,682] Trial 7 finished with value: 0.6562905644151281 and parameters: {'num_leaves': 10}. Best is trial 7 with value: 0.6562905644151281. num_leaves, val_score: 0.655992: 5%|5 | 1/20 [00:00<00:07, 2.47it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000313 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.313854 valid's binary_logloss: 0.672207 Early stopping, best iteration is: [34] train's binary_logloss: 0.498751 valid's binary_logloss: 0.663462 num_leaves, val_score: 0.655992: 10%|# | 2/20 [00:01<00:13, 1.32it/s][I 2020-09-27 04:39:17,270] Trial 8 finished with value: 0.6634620451568378 and parameters: {'num_leaves': 238}. Best is trial 7 with value: 0.6562905644151281. num_leaves, val_score: 0.655992: 10%|# | 2/20 [00:02<00:13, 1.32it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000229 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.442275 valid's binary_logloss: 0.665841 Early stopping, best iteration is: [40] train's binary_logloss: 0.552251 valid's binary_logloss: 0.660374 num_leaves, val_score: 0.655992: 15%|#5 | 3/20 [00:02<00:12, 1.37it/s][I 2020-09-27 04:39:17,923] Trial 9 finished with value: 0.6603744399138358 and parameters: {'num_leaves': 116}. Best is trial 7 with value: 0.6562905644151281. num_leaves, val_score: 0.655992: 15%|#5 | 3/20 [00:02<00:12, 1.37it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010332 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.308823 valid's binary_logloss: 0.670809 Early stopping, best iteration is: [36] train's binary_logloss: 0.48791 valid's binary_logloss: 0.665738 num_leaves, val_score: 0.655992: 20%|## | 4/20 [00:03<00:12, 1.27it/s][I 2020-09-27 04:39:18,840] Trial 10 finished with value: 0.6657380834337528 and parameters: {'num_leaves': 244}. Best is trial 7 with value: 0.6562905644151281. num_leaves, val_score: 0.655992: 20%|## | 4/20 [00:03<00:12, 1.27it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000322 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.658235 valid's binary_logloss: 0.664731 [200] train's binary_logloss: 0.647762 valid's binary_logloss: 0.657647 [300] train's binary_logloss: 0.642542 valid's binary_logloss: 0.655465 [400] train's binary_logloss: 0.638941 valid's binary_logloss: 0.654486 [500] train's binary_logloss: 0.636035 valid's binary_logloss: 0.653981 [600] train's binary_logloss: 0.63343 valid's binary_logloss: 0.65371 Early stopping, best iteration is: [555] train's binary_logloss: 0.634597 valid's binary_logloss: 0.653553 num_leaves, val_score: 0.653553: 25%|##5 | 5/20 [00:04<00:11, 1.26it/s][I 2020-09-27 04:39:19,645] Trial 11 finished with value: 0.6535525819216924 and parameters: {'num_leaves': 3}. Best is trial 11 with value: 0.6535525819216924. num_leaves, val_score: 0.653553: 25%|##5 | 5/20 [00:04<00:11, 1.26it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000332 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.620698 valid's binary_logloss: 0.657525 [200] train's binary_logloss: 0.594023 valid's binary_logloss: 0.657636 Early stopping, best iteration is: [151] train's binary_logloss: 0.606458 valid's binary_logloss: 0.657287 num_leaves, val_score: 0.653553: 30%|### | 6/20 [00:05<00:10, 1.32it/s][I 2020-09-27 04:39:20,325] Trial 12 finished with value: 0.6572874929871441 and parameters: {'num_leaves': 14}. Best is trial 11 with value: 0.6535525819216924. num_leaves, val_score: 0.653553: 30%|### | 6/20 [00:05<00:10, 1.32it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011876 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.467946 valid's binary_logloss: 0.670734 Early stopping, best iteration is: [45] train's binary_logloss: 0.556356 valid's binary_logloss: 0.664088 num_leaves, val_score: 0.653553: 35%|###5 | 7/20 [00:05<00:10, 1.25it/s][I 2020-09-27 04:39:21,228] Trial 13 finished with value: 0.6640878662544877 and parameters: {'num_leaves': 97}. Best is trial 11 with value: 0.6535525819216924. num_leaves, val_score: 0.653553: 35%|###5 | 7/20 [00:05<00:10, 1.25it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000489 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.524406 valid's binary_logloss: 0.663032 Early stopping, best iteration is: [41] train's binary_logloss: 0.595701 valid's binary_logloss: 0.662191 num_leaves, val_score: 0.653553: 40%|#### | 8/20 [00:06<00:08, 1.39it/s][I 2020-09-27 04:39:21,759] Trial 14 finished with value: 0.6621905066342226 and parameters: {'num_leaves': 62}. Best is trial 11 with value: 0.6535525819216924. num_leaves, val_score: 0.653553: 40%|#### | 8/20 [00:06<00:08, 1.39it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000238 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.367271 valid's binary_logloss: 0.673756 Early stopping, best iteration is: [34] train's binary_logloss: 0.528584 valid's binary_logloss: 0.666502 num_leaves, val_score: 0.653553: 45%|####5 | 9/20 [00:07<00:08, 1.30it/s][I 2020-09-27 04:39:22,638] Trial 15 finished with value: 0.6665016344232115 and parameters: {'num_leaves': 180}. Best is trial 11 with value: 0.6535525819216924. num_leaves, val_score: 0.653553: 45%|####5 | 9/20 [00:07<00:08, 1.30it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000232 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.36627 valid's binary_logloss: 0.674896 Early stopping, best iteration is: [36] train's binary_logloss: 0.521716 valid's binary_logloss: 0.66343 num_leaves, val_score: 0.653553: 50%|##### | 10/20 [00:08<00:08, 1.21it/s][I 2020-09-27 04:39:23,590] Trial 16 finished with value: 0.6634300540848975 and parameters: {'num_leaves': 183}. Best is trial 11 with value: 0.6535525819216924. num_leaves, val_score: 0.653553: 50%|##### | 10/20 [00:08<00:08, 1.21it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000415 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.536123 valid's binary_logloss: 0.658196 Early stopping, best iteration is: [92] train's binary_logloss: 0.543103 valid's binary_logloss: 0.657473 num_leaves, val_score: 0.653553: 55%|#####5 | 11/20 [00:09<00:07, 1.13it/s][I 2020-09-27 04:39:24,613] Trial 17 finished with value: 0.6574725321803361 and parameters: {'num_leaves': 55}. Best is trial 11 with value: 0.6535525819216924. num_leaves, val_score: 0.653553: 55%|#####5 | 11/20 [00:09<00:07, 1.13it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000386 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.386366 valid's binary_logloss: 0.670881 Early stopping, best iteration is: [32] train's binary_logloss: 0.54618 valid's binary_logloss: 0.663828 num_leaves, val_score: 0.653553: 60%|###### | 12/20 [00:10<00:07, 1.14it/s][I 2020-09-27 04:39:25,475] Trial 18 finished with value: 0.663827571242285 and parameters: {'num_leaves': 162}. Best is trial 11 with value: 0.6535525819216924. num_leaves, val_score: 0.653553: 60%|###### | 12/20 [00:10<00:07, 1.14it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000380 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.543127 valid's binary_logloss: 0.660244 Early stopping, best iteration is: [92] train's binary_logloss: 0.549667 valid's binary_logloss: 0.659503 num_leaves, val_score: 0.653553: 65%|######5 | 13/20 [00:10<00:05, 1.28it/s][I 2020-09-27 04:39:26,025] Trial 19 finished with value: 0.6595031085541457 and parameters: {'num_leaves': 51}. Best is trial 11 with value: 0.6535525819216924. num_leaves, val_score: 0.653553: 65%|######5 | 13/20 [00:10<00:05, 1.28it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000521 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.636989 valid's binary_logloss: 0.657697 [200] train's binary_logloss: 0.620568 valid's binary_logloss: 0.656088 Early stopping, best iteration is: [171] train's binary_logloss: 0.624795 valid's binary_logloss: 0.655485 num_leaves, val_score: 0.653553: 70%|####### | 14/20 [00:11<00:04, 1.49it/s][I 2020-09-27 04:39:26,451] Trial 20 finished with value: 0.6554852556407339 and parameters: {'num_leaves': 8}. Best is trial 11 with value: 0.6535525819216924. num_leaves, val_score: 0.653553: 70%|####### | 14/20 [00:11<00:04, 1.49it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004552 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.643142 valid's binary_logloss: 0.658677 [200] train's binary_logloss: 0.62998 valid's binary_logloss: 0.656522 [300] train's binary_logloss: 0.620354 valid's binary_logloss: 0.655925 Early stopping, best iteration is: [297] train's binary_logloss: 0.620584 valid's binary_logloss: 0.655776 num_leaves, val_score: 0.653553: 75%|#######5 | 15/20 [00:11<00:03, 1.58it/s][I 2020-09-27 04:39:26,995] Trial 21 finished with value: 0.655776407729715 and parameters: {'num_leaves': 6}. Best is trial 11 with value: 0.6535525819216924. num_leaves, val_score: 0.653553: 75%|#######5 | 15/20 [00:11<00:03, 1.58it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000245 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.639989 valid's binary_logloss: 0.657478 [200] train's binary_logloss: 0.62545 valid's binary_logloss: 0.65644 Early stopping, best iteration is: [140] train's binary_logloss: 0.633249 valid's binary_logloss: 0.655769 num_leaves, val_score: 0.653553: 80%|######## | 16/20 [00:12<00:02, 1.77it/s][I 2020-09-27 04:39:27,399] Trial 22 finished with value: 0.6557689832038824 and parameters: {'num_leaves': 7}. Best is trial 11 with value: 0.6535525819216924. num_leaves, val_score: 0.653553: 80%|######## | 16/20 [00:12<00:02, 1.77it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004511 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.575815 valid's binary_logloss: 0.658024 [200] train's binary_logloss: 0.522157 valid's binary_logloss: 0.657932 Early stopping, best iteration is: [154] train's binary_logloss: 0.544614 valid's binary_logloss: 0.656365 num_leaves, val_score: 0.653553: 85%|########5 | 17/20 [00:13<00:02, 1.40it/s][I 2020-09-27 04:39:28,462] Trial 23 finished with value: 0.6563652624863247 and parameters: {'num_leaves': 34}. Best is trial 11 with value: 0.6535525819216924. num_leaves, val_score: 0.653553: 85%|########5 | 17/20 [00:13<00:02, 1.40it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009938 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.647155 valid's binary_logloss: 0.660323 [200] train's binary_logloss: 0.635295 valid's binary_logloss: 0.655924 Early stopping, best iteration is: [188] train's binary_logloss: 0.63642 valid's binary_logloss: 0.655643 num_leaves, val_score: 0.653553: 90%|######### | 18/20 [00:13<00:01, 1.59it/s][I 2020-09-27 04:39:28,891] Trial 24 finished with value: 0.655643287434117 and parameters: {'num_leaves': 5}. Best is trial 11 with value: 0.6535525819216924. num_leaves, val_score: 0.653553: 90%|######### | 18/20 [00:13<00:01, 1.59it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000491 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.488408 valid's binary_logloss: 0.664489 Early stopping, best iteration is: [78] train's binary_logloss: 0.517222 valid's binary_logloss: 0.661867 num_leaves, val_score: 0.653553: 95%|#########5| 19/20 [00:14<00:00, 1.53it/s][I 2020-09-27 04:39:29,600] Trial 25 finished with value: 0.6618666948452248 and parameters: {'num_leaves': 84}. Best is trial 11 with value: 0.6535525819216924. num_leaves, val_score: 0.653553: 95%|#########5| 19/20 [00:14<00:00, 1.53it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000236 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.580318 valid's binary_logloss: 0.659611 Early stopping, best iteration is: [67] train's binary_logloss: 0.601799 valid's binary_logloss: 0.658994 num_leaves, val_score: 0.653553: 100%|##########| 20/20 [00:14<00:00, 1.71it/s][I 2020-09-27 04:39:30,031] Trial 26 finished with value: 0.6589942104081719 and parameters: {'num_leaves': 32}. Best is trial 11 with value: 0.6535525819216924. num_leaves, val_score: 0.653553: 100%|##########| 20/20 [00:14<00:00, 1.35it/s] bagging, val_score: 0.653553: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.020148 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657525 valid's binary_logloss: 0.662893 [200] train's binary_logloss: 0.647047 valid's binary_logloss: 0.656095 [300] train's binary_logloss: 0.641935 valid's binary_logloss: 0.654303 [400] train's binary_logloss: 0.638294 valid's binary_logloss: 0.653695 [500] train's binary_logloss: 0.635159 valid's binary_logloss: 0.653205 [600] train's binary_logloss: 0.632193 valid's binary_logloss: 0.652803 [700] train's binary_logloss: 0.629395 valid's binary_logloss: 0.651799 [800] train's binary_logloss: 0.626828 valid's binary_logloss: 0.652063 Early stopping, best iteration is: [745] train's binary_logloss: 0.628188 valid's binary_logloss: 0.651395 bagging, val_score: 0.651395: 10%|# | 1/10 [00:01<00:10, 1.19s/it][I 2020-09-27 04:39:31,244] Trial 27 finished with value: 0.6513952391116833 and parameters: {'bagging_fraction': 0.8097719374771206, 'bagging_freq': 2}. Best is trial 27 with value: 0.6513952391116833. bagging, val_score: 0.651395: 10%|# | 1/10 [00:01<00:10, 1.19s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000403 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657637 valid's binary_logloss: 0.663101 [200] train's binary_logloss: 0.647154 valid's binary_logloss: 0.656096 [300] train's binary_logloss: 0.642211 valid's binary_logloss: 0.653702 [400] train's binary_logloss: 0.638597 valid's binary_logloss: 0.652862 [500] train's binary_logloss: 0.635489 valid's binary_logloss: 0.652354 [600] train's binary_logloss: 0.632512 valid's binary_logloss: 0.652123 [700] train's binary_logloss: 0.62981 valid's binary_logloss: 0.651852 [800] train's binary_logloss: 0.627226 valid's binary_logloss: 0.651697 Early stopping, best iteration is: [743] train's binary_logloss: 0.628657 valid's binary_logloss: 0.651255 bagging, val_score: 0.651255: 20%|## | 2/10 [00:02<00:10, 1.35s/it][I 2020-09-27 04:39:32,959] Trial 28 finished with value: 0.6512546217026939 and parameters: {'bagging_fraction': 0.8280339557117807, 'bagging_freq': 2}. Best is trial 28 with value: 0.6512546217026939. bagging, val_score: 0.651255: 20%|## | 2/10 [00:02<00:10, 1.35s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000249 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657845 valid's binary_logloss: 0.663325 [200] train's binary_logloss: 0.647265 valid's binary_logloss: 0.656205 [300] train's binary_logloss: 0.642195 valid's binary_logloss: 0.653918 [400] train's binary_logloss: 0.638463 valid's binary_logloss: 0.653807 [500] train's binary_logloss: 0.635325 valid's binary_logloss: 0.653223 [600] train's binary_logloss: 0.632547 valid's binary_logloss: 0.652408 [700] train's binary_logloss: 0.629962 valid's binary_logloss: 0.652142 [800] train's binary_logloss: 0.627395 valid's binary_logloss: 0.652339 Early stopping, best iteration is: [743] train's binary_logloss: 0.628893 valid's binary_logloss: 0.651858 bagging, val_score: 0.651255: 30%|### | 3/10 [00:04<00:09, 1.32s/it][I 2020-09-27 04:39:34,223] Trial 29 finished with value: 0.6518582592610941 and parameters: {'bagging_fraction': 0.8423665601800546, 'bagging_freq': 2}. Best is trial 28 with value: 0.6512546217026939. bagging, val_score: 0.651255: 30%|### | 3/10 [00:04<00:09, 1.32s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000300 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657643 valid's binary_logloss: 0.663425 [200] train's binary_logloss: 0.647094 valid's binary_logloss: 0.65631 [300] train's binary_logloss: 0.642016 valid's binary_logloss: 0.653742 [400] train's binary_logloss: 0.638407 valid's binary_logloss: 0.653279 Early stopping, best iteration is: [377] train's binary_logloss: 0.639168 valid's binary_logloss: 0.653138 bagging, val_score: 0.651255: 40%|#### | 4/10 [00:04<00:06, 1.14s/it][I 2020-09-27 04:39:34,939] Trial 30 finished with value: 0.6531380481905572 and parameters: {'bagging_fraction': 0.8282561902576878, 'bagging_freq': 2}. Best is trial 28 with value: 0.6512546217026939. bagging, val_score: 0.651255: 40%|#### | 4/10 [00:04<00:06, 1.14s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000244 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.65787 valid's binary_logloss: 0.662913 [200] train's binary_logloss: 0.647202 valid's binary_logloss: 0.656389 [300] train's binary_logloss: 0.642181 valid's binary_logloss: 0.65432 [400] train's binary_logloss: 0.638556 valid's binary_logloss: 0.653816 Early stopping, best iteration is: [361] train's binary_logloss: 0.639969 valid's binary_logloss: 0.653683 bagging, val_score: 0.651255: 50%|##### | 5/10 [00:06<00:05, 1.14s/it][I 2020-09-27 04:39:36,081] Trial 31 finished with value: 0.6536825951065313 and parameters: {'bagging_fraction': 0.8381544500351495, 'bagging_freq': 2}. Best is trial 28 with value: 0.6512546217026939. bagging, val_score: 0.651255: 50%|##### | 5/10 [00:06<00:05, 1.14s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000455 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657707 valid's binary_logloss: 0.662425 [200] train's binary_logloss: 0.64726 valid's binary_logloss: 0.655364 [300] train's binary_logloss: 0.64231 valid's binary_logloss: 0.653736 [400] train's binary_logloss: 0.638629 valid's binary_logloss: 0.653251 [500] train's binary_logloss: 0.63557 valid's binary_logloss: 0.652996 [600] train's binary_logloss: 0.632836 valid's binary_logloss: 0.653094 [700] train's binary_logloss: 0.630172 valid's binary_logloss: 0.652558 [800] train's binary_logloss: 0.627493 valid's binary_logloss: 0.652618 Early stopping, best iteration is: [738] train's binary_logloss: 0.629206 valid's binary_logloss: 0.65221 bagging, val_score: 0.651255: 60%|###### | 6/10 [00:07<00:04, 1.16s/it][I 2020-09-27 04:39:37,277] Trial 32 finished with value: 0.6522103084600892 and parameters: {'bagging_fraction': 0.8240973522704712, 'bagging_freq': 2}. Best is trial 28 with value: 0.6512546217026939. bagging, val_score: 0.651255: 60%|###### | 6/10 [00:07<00:04, 1.16s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000240 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657759 valid's binary_logloss: 0.6633 [200] train's binary_logloss: 0.647172 valid's binary_logloss: 0.656277 [300] train's binary_logloss: 0.642123 valid's binary_logloss: 0.654052 [400] train's binary_logloss: 0.638549 valid's binary_logloss: 0.653547 [500] train's binary_logloss: 0.635603 valid's binary_logloss: 0.653766 Early stopping, best iteration is: [419] train's binary_logloss: 0.637971 valid's binary_logloss: 0.653468 bagging, val_score: 0.651255: 70%|####### | 7/10 [00:08<00:03, 1.04s/it][I 2020-09-27 04:39:38,055] Trial 33 finished with value: 0.6534681190347887 and parameters: {'bagging_fraction': 0.8133668384521671, 'bagging_freq': 2}. Best is trial 28 with value: 0.6512546217026939. bagging, val_score: 0.651255: 70%|####### | 7/10 [00:08<00:03, 1.04s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000309 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.65767 valid's binary_logloss: 0.663092 [200] train's binary_logloss: 0.646954 valid's binary_logloss: 0.65622 [300] train's binary_logloss: 0.641882 valid's binary_logloss: 0.653856 [400] train's binary_logloss: 0.638359 valid's binary_logloss: 0.653238 [500] train's binary_logloss: 0.635264 valid's binary_logloss: 0.652485 [600] train's binary_logloss: 0.632569 valid's binary_logloss: 0.652572 Early stopping, best iteration is: [500] train's binary_logloss: 0.635264 valid's binary_logloss: 0.652485 bagging, val_score: 0.651255: 80%|######## | 8/10 [00:08<00:01, 1.01it/s][I 2020-09-27 04:39:38,927] Trial 34 finished with value: 0.6524845330402806 and parameters: {'bagging_fraction': 0.833189344448115, 'bagging_freq': 2}. Best is trial 28 with value: 0.6512546217026939. bagging, val_score: 0.651255: 80%|######## | 8/10 [00:08<00:01, 1.01it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000447 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657357 valid's binary_logloss: 0.663082 [200] train's binary_logloss: 0.646962 valid's binary_logloss: 0.656317 [300] train's binary_logloss: 0.642048 valid's binary_logloss: 0.65432 Early stopping, best iteration is: [285] train's binary_logloss: 0.642706 valid's binary_logloss: 0.653911 bagging, val_score: 0.651255: 90%|######### | 9/10 [00:09<00:01, 1.02s/it][I 2020-09-27 04:39:40,016] Trial 35 finished with value: 0.653910601498707 and parameters: {'bagging_fraction': 0.6953833757673704, 'bagging_freq': 3}. Best is trial 28 with value: 0.6512546217026939. bagging, val_score: 0.651255: 90%|######### | 9/10 [00:09<00:01, 1.02s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000404 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.658127 valid's binary_logloss: 0.664339 [200] train's binary_logloss: 0.647507 valid's binary_logloss: 0.656812 [300] train's binary_logloss: 0.642315 valid's binary_logloss: 0.65461 [400] train's binary_logloss: 0.638734 valid's binary_logloss: 0.653365 [500] train's binary_logloss: 0.635615 valid's binary_logloss: 0.653259 Early stopping, best iteration is: [439] train's binary_logloss: 0.637483 valid's binary_logloss: 0.652984 bagging, val_score: 0.651255: 100%|##########| 10/10 [00:10<00:00, 1.03it/s][I 2020-09-27 04:39:40,868] Trial 36 finished with value: 0.6529843808564952 and parameters: {'bagging_fraction': 0.9750898583135423, 'bagging_freq': 5}. Best is trial 28 with value: 0.6512546217026939. bagging, val_score: 0.651255: 100%|##########| 10/10 [00:10<00:00, 1.08s/it] feature_fraction_stage2, val_score: 0.651255: 0%| | 0/3 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000412 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657426 valid's binary_logloss: 0.66294 [200] train's binary_logloss: 0.646763 valid's binary_logloss: 0.65579 [300] train's binary_logloss: 0.641842 valid's binary_logloss: 0.653608 [400] train's binary_logloss: 0.638277 valid's binary_logloss: 0.653005 [500] train's binary_logloss: 0.635065 valid's binary_logloss: 0.652871 Early stopping, best iteration is: [437] train's binary_logloss: 0.637058 valid's binary_logloss: 0.652763 feature_fraction_stage2, val_score: 0.651255: 33%|###3 | 1/3 [00:00<00:01, 1.20it/s][I 2020-09-27 04:39:41,719] Trial 37 finished with value: 0.6527633350213652 and parameters: {'feature_fraction': 0.41600000000000004}. Best is trial 37 with value: 0.6527633350213652. feature_fraction_stage2, val_score: 0.651255: 33%|###3 | 1/3 [00:00<00:01, 1.20it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000510 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657376 valid's binary_logloss: 0.66298 [200] train's binary_logloss: 0.646944 valid's binary_logloss: 0.656526 [300] train's binary_logloss: 0.641961 valid's binary_logloss: 0.653998 [400] train's binary_logloss: 0.638308 valid's binary_logloss: 0.653033 [500] train's binary_logloss: 0.634963 valid's binary_logloss: 0.652584 [600] train's binary_logloss: 0.632079 valid's binary_logloss: 0.652716 [700] train's binary_logloss: 0.629142 valid's binary_logloss: 0.652471 [800] train's binary_logloss: 0.62641 valid's binary_logloss: 0.652186 Early stopping, best iteration is: [758] train's binary_logloss: 0.627519 valid's binary_logloss: 0.651784 feature_fraction_stage2, val_score: 0.651255: 67%|######6 | 2/3 [00:02<00:00, 1.00it/s][I 2020-09-27 04:39:43,108] Trial 38 finished with value: 0.6517839548740617 and parameters: {'feature_fraction': 0.44800000000000006}. Best is trial 38 with value: 0.6517839548740617. feature_fraction_stage2, val_score: 0.651255: 67%|######6 | 2/3 [00:02<00:00, 1.00it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011259 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657376 valid's binary_logloss: 0.66298 [200] train's binary_logloss: 0.646944 valid's binary_logloss: 0.656526 [300] train's binary_logloss: 0.641961 valid's binary_logloss: 0.653998 [400] train's binary_logloss: 0.638308 valid's binary_logloss: 0.653033 [500] train's binary_logloss: 0.634963 valid's binary_logloss: 0.652584 [600] train's binary_logloss: 0.632079 valid's binary_logloss: 0.652716 [700] train's binary_logloss: 0.629142 valid's binary_logloss: 0.652471 [800] train's binary_logloss: 0.62641 valid's binary_logloss: 0.652186 Early stopping, best iteration is: [758] train's binary_logloss: 0.627519 valid's binary_logloss: 0.651784 feature_fraction_stage2, val_score: 0.651255: 100%|##########| 3/3 [00:03<00:00, 1.15s/it][I 2020-09-27 04:39:44,601] Trial 39 finished with value: 0.6517839548740617 and parameters: {'feature_fraction': 0.48000000000000004}. Best is trial 38 with value: 0.6517839548740617. feature_fraction_stage2, val_score: 0.651255: 100%|##########| 3/3 [00:03<00:00, 1.24s/it] regularization_factors, val_score: 0.651255: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000460 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657637 valid's binary_logloss: 0.663101 [200] train's binary_logloss: 0.647154 valid's binary_logloss: 0.656096 [300] train's binary_logloss: 0.642211 valid's binary_logloss: 0.653702 [400] train's binary_logloss: 0.638597 valid's binary_logloss: 0.652862 [500] train's binary_logloss: 0.635489 valid's binary_logloss: 0.652354 [600] train's binary_logloss: 0.632512 valid's binary_logloss: 0.652123 [700] train's binary_logloss: 0.62981 valid's binary_logloss: 0.651852 [800] train's binary_logloss: 0.627227 valid's binary_logloss: 0.651697 Early stopping, best iteration is: [743] train's binary_logloss: 0.628658 valid's binary_logloss: 0.651255 regularization_factors, val_score: 0.651255: 5%|5 | 1/20 [00:01<00:23, 1.23s/it][I 2020-09-27 04:39:45,848] Trial 40 finished with value: 0.6512546032948433 and parameters: {'lambda_l1': 3.7407304516474708e-06, 'lambda_l2': 0.0003236321792160619}. Best is trial 40 with value: 0.6512546032948433. regularization_factors, val_score: 0.651255: 5%|5 | 1/20 [00:01<00:23, 1.23s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000479 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657637 valid's binary_logloss: 0.663101 [200] train's binary_logloss: 0.647154 valid's binary_logloss: 0.656096 [300] train's binary_logloss: 0.642211 valid's binary_logloss: 0.653702 [400] train's binary_logloss: 0.638597 valid's binary_logloss: 0.652862 [500] train's binary_logloss: 0.635489 valid's binary_logloss: 0.652354 [600] train's binary_logloss: 0.632512 valid's binary_logloss: 0.652123 [700] train's binary_logloss: 0.62981 valid's binary_logloss: 0.651852 [800] train's binary_logloss: 0.627227 valid's binary_logloss: 0.651697 Early stopping, best iteration is: [743] train's binary_logloss: 0.628658 valid's binary_logloss: 0.651255 regularization_factors, val_score: 0.651255: 10%|# | 2/20 [00:02<00:23, 1.33s/it][I 2020-09-27 04:39:47,419] Trial 41 finished with value: 0.6512546004999574 and parameters: {'lambda_l1': 1.58423580909119e-06, 'lambda_l2': 0.00037329518153918976}. Best is trial 41 with value: 0.6512546004999574. regularization_factors, val_score: 0.651255: 10%|# | 2/20 [00:02<00:23, 1.33s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000390 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657637 valid's binary_logloss: 0.663101 [200] train's binary_logloss: 0.647154 valid's binary_logloss: 0.656096 [300] train's binary_logloss: 0.642211 valid's binary_logloss: 0.653702 [400] train's binary_logloss: 0.638597 valid's binary_logloss: 0.652862 [500] train's binary_logloss: 0.635489 valid's binary_logloss: 0.652354 [600] train's binary_logloss: 0.632512 valid's binary_logloss: 0.652123 [700] train's binary_logloss: 0.62981 valid's binary_logloss: 0.651852 [800] train's binary_logloss: 0.627227 valid's binary_logloss: 0.651697 Early stopping, best iteration is: [743] train's binary_logloss: 0.628658 valid's binary_logloss: 0.651255 regularization_factors, val_score: 0.651255: 15%|#5 | 3/20 [00:04<00:22, 1.33s/it][I 2020-09-27 04:39:48,734] Trial 42 finished with value: 0.6512545924218361 and parameters: {'lambda_l1': 1.3626757884196743e-06, 'lambda_l2': 0.0005160151192105561}. Best is trial 42 with value: 0.6512545924218361. regularization_factors, val_score: 0.651255: 15%|#5 | 3/20 [00:04<00:22, 1.33s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000381 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657637 valid's binary_logloss: 0.663101 [200] train's binary_logloss: 0.647154 valid's binary_logloss: 0.656096 [300] train's binary_logloss: 0.642211 valid's binary_logloss: 0.653702 [400] train's binary_logloss: 0.638597 valid's binary_logloss: 0.652862 [500] train's binary_logloss: 0.635489 valid's binary_logloss: 0.652354 [600] train's binary_logloss: 0.632512 valid's binary_logloss: 0.652123 [700] train's binary_logloss: 0.62981 valid's binary_logloss: 0.651852 [800] train's binary_logloss: 0.627227 valid's binary_logloss: 0.651697 Early stopping, best iteration is: [743] train's binary_logloss: 0.628658 valid's binary_logloss: 0.651255 regularization_factors, val_score: 0.651255: 20%|## | 4/20 [00:05<00:20, 1.29s/it][I 2020-09-27 04:39:49,939] Trial 43 finished with value: 0.6512546024195541 and parameters: {'lambda_l1': 1.826384761461434e-06, 'lambda_l2': 0.000339404582931327}. Best is trial 42 with value: 0.6512545924218361. regularization_factors, val_score: 0.651255: 20%|## | 4/20 [00:05<00:20, 1.29s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000442 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657637 valid's binary_logloss: 0.663101 [200] train's binary_logloss: 0.647154 valid's binary_logloss: 0.656096 [300] train's binary_logloss: 0.642211 valid's binary_logloss: 0.653702 [400] train's binary_logloss: 0.638597 valid's binary_logloss: 0.652862 [500] train's binary_logloss: 0.635489 valid's binary_logloss: 0.652354 [600] train's binary_logloss: 0.632512 valid's binary_logloss: 0.652123 [700] train's binary_logloss: 0.62981 valid's binary_logloss: 0.651852 [800] train's binary_logloss: 0.627227 valid's binary_logloss: 0.651697 Early stopping, best iteration is: [743] train's binary_logloss: 0.628658 valid's binary_logloss: 0.651255 regularization_factors, val_score: 0.651255: 25%|##5 | 5/20 [00:07<00:21, 1.42s/it][I 2020-09-27 04:39:51,672] Trial 44 finished with value: 0.6512545971181739 and parameters: {'lambda_l1': 9.830713052753886e-07, 'lambda_l2': 0.00043290604666443164}. Best is trial 42 with value: 0.6512545924218361. regularization_factors, val_score: 0.651255: 25%|##5 | 5/20 [00:07<00:21, 1.42s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000564 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657637 valid's binary_logloss: 0.663101 [200] train's binary_logloss: 0.647154 valid's binary_logloss: 0.656096 [300] train's binary_logloss: 0.642211 valid's binary_logloss: 0.653702 [400] train's binary_logloss: 0.638597 valid's binary_logloss: 0.652862 [500] train's binary_logloss: 0.635489 valid's binary_logloss: 0.652354 [600] train's binary_logloss: 0.632512 valid's binary_logloss: 0.652123 [700] train's binary_logloss: 0.62981 valid's binary_logloss: 0.651852 [800] train's binary_logloss: 0.627227 valid's binary_logloss: 0.651697 Early stopping, best iteration is: [743] train's binary_logloss: 0.628658 valid's binary_logloss: 0.651255 regularization_factors, val_score: 0.651255: 30%|### | 6/20 [00:08<00:19, 1.36s/it][I 2020-09-27 04:39:52,891] Trial 45 finished with value: 0.6512546006093322 and parameters: {'lambda_l1': 7.770048321005081e-07, 'lambda_l2': 0.00037155770796455103}. Best is trial 42 with value: 0.6512545924218361. regularization_factors, val_score: 0.651255: 30%|### | 6/20 [00:08<00:19, 1.36s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000386 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657637 valid's binary_logloss: 0.663101 [200] train's binary_logloss: 0.647154 valid's binary_logloss: 0.656096 [300] train's binary_logloss: 0.642211 valid's binary_logloss: 0.653702 [400] train's binary_logloss: 0.638597 valid's binary_logloss: 0.652862 [500] train's binary_logloss: 0.635489 valid's binary_logloss: 0.652354 [600] train's binary_logloss: 0.632512 valid's binary_logloss: 0.652123 [700] train's binary_logloss: 0.62981 valid's binary_logloss: 0.651852 [800] train's binary_logloss: 0.627227 valid's binary_logloss: 0.651697 Early stopping, best iteration is: [743] train's binary_logloss: 0.628658 valid's binary_logloss: 0.651255 regularization_factors, val_score: 0.651255: 35%|###5 | 7/20 [00:09<00:17, 1.32s/it][I 2020-09-27 04:39:54,130] Trial 46 finished with value: 0.65125458769479 and parameters: {'lambda_l1': 5.65299654936304e-07, 'lambda_l2': 0.0005992895040580498}. Best is trial 46 with value: 0.65125458769479. regularization_factors, val_score: 0.651255: 35%|###5 | 7/20 [00:09<00:17, 1.32s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000244 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657637 valid's binary_logloss: 0.663101 [200] train's binary_logloss: 0.647154 valid's binary_logloss: 0.656096 [300] train's binary_logloss: 0.642211 valid's binary_logloss: 0.653702 [400] train's binary_logloss: 0.638597 valid's binary_logloss: 0.652862 [500] train's binary_logloss: 0.635489 valid's binary_logloss: 0.652354 [600] train's binary_logloss: 0.632512 valid's binary_logloss: 0.652123 [700] train's binary_logloss: 0.62981 valid's binary_logloss: 0.651852 [800] train's binary_logloss: 0.627227 valid's binary_logloss: 0.651697 Early stopping, best iteration is: [743] train's binary_logloss: 0.628658 valid's binary_logloss: 0.651255 regularization_factors, val_score: 0.651255: 40%|#### | 8/20 [00:11<00:17, 1.42s/it][I 2020-09-27 04:39:55,765] Trial 47 finished with value: 0.651254582843981 and parameters: {'lambda_l1': 5.113147117976975e-07, 'lambda_l2': 0.0006849316707389379}. Best is trial 47 with value: 0.651254582843981. regularization_factors, val_score: 0.651255: 40%|#### | 8/20 [00:11<00:17, 1.42s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000452 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657637 valid's binary_logloss: 0.663101 [200] train's binary_logloss: 0.647154 valid's binary_logloss: 0.656096 [300] train's binary_logloss: 0.642211 valid's binary_logloss: 0.653702 [400] train's binary_logloss: 0.638597 valid's binary_logloss: 0.652862 [500] train's binary_logloss: 0.635489 valid's binary_logloss: 0.652354 [600] train's binary_logloss: 0.632512 valid's binary_logloss: 0.652123 [700] train's binary_logloss: 0.62981 valid's binary_logloss: 0.651852 [800] train's binary_logloss: 0.627227 valid's binary_logloss: 0.651697 Early stopping, best iteration is: [743] train's binary_logloss: 0.628658 valid's binary_logloss: 0.651255 regularization_factors, val_score: 0.651255: 45%|####5 | 9/20 [00:12<00:14, 1.35s/it][I 2020-09-27 04:39:56,969] Trial 48 finished with value: 0.6512545741011545 and parameters: {'lambda_l1': 3.656814382369128e-07, 'lambda_l2': 0.0008392662361736358}. Best is trial 48 with value: 0.6512545741011545. regularization_factors, val_score: 0.651255: 45%|####5 | 9/20 [00:12<00:14, 1.35s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000266 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657637 valid's binary_logloss: 0.663101 [200] train's binary_logloss: 0.647154 valid's binary_logloss: 0.656096 [300] train's binary_logloss: 0.642211 valid's binary_logloss: 0.653702 [400] train's binary_logloss: 0.638597 valid's binary_logloss: 0.652862 [500] train's binary_logloss: 0.635489 valid's binary_logloss: 0.652354 [600] train's binary_logloss: 0.632513 valid's binary_logloss: 0.652123 [700] train's binary_logloss: 0.629811 valid's binary_logloss: 0.651852 [800] train's binary_logloss: 0.627228 valid's binary_logloss: 0.651696 Early stopping, best iteration is: [743] train's binary_logloss: 0.628659 valid's binary_logloss: 0.651254 regularization_factors, val_score: 0.651254: 50%|##### | 10/20 [00:13<00:13, 1.31s/it][I 2020-09-27 04:39:58,184] Trial 49 finished with value: 0.65125447186593 and parameters: {'lambda_l1': 4.097168681166982e-08, 'lambda_l2': 0.002642450368328453}. Best is trial 49 with value: 0.65125447186593. regularization_factors, val_score: 0.651254: 50%|##### | 10/20 [00:13<00:13, 1.31s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000234 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657642 valid's binary_logloss: 0.663104 [200] train's binary_logloss: 0.647161 valid's binary_logloss: 0.6561 [300] train's binary_logloss: 0.642211 valid's binary_logloss: 0.653789 [400] train's binary_logloss: 0.638567 valid's binary_logloss: 0.652995 Early stopping, best iteration is: [378] train's binary_logloss: 0.639273 valid's binary_logloss: 0.65285 regularization_factors, val_score: 0.651254: 55%|#####5 | 11/20 [00:14<00:11, 1.26s/it][I 2020-09-27 04:39:59,310] Trial 50 finished with value: 0.6528496531447342 and parameters: {'lambda_l1': 1.8926948433638843e-08, 'lambda_l2': 0.18218516904146864}. Best is trial 49 with value: 0.65125447186593. regularization_factors, val_score: 0.651254: 55%|#####5 | 11/20 [00:14<00:11, 1.26s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000499 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657637 valid's binary_logloss: 0.663101 [200] train's binary_logloss: 0.647154 valid's binary_logloss: 0.656096 [300] train's binary_logloss: 0.642211 valid's binary_logloss: 0.653702 [400] train's binary_logloss: 0.638597 valid's binary_logloss: 0.652862 [500] train's binary_logloss: 0.635489 valid's binary_logloss: 0.652354 [600] train's binary_logloss: 0.632513 valid's binary_logloss: 0.652123 [700] train's binary_logloss: 0.629811 valid's binary_logloss: 0.651852 [800] train's binary_logloss: 0.627228 valid's binary_logloss: 0.651697 Early stopping, best iteration is: [743] train's binary_logloss: 0.628659 valid's binary_logloss: 0.651254 regularization_factors, val_score: 0.651254: 60%|###### | 12/20 [00:15<00:10, 1.25s/it][I 2020-09-27 04:40:00,559] Trial 51 finished with value: 0.651254491688544 and parameters: {'lambda_l1': 9.464855403688449e-08, 'lambda_l2': 0.0022922152533675836}. Best is trial 49 with value: 0.65125447186593. regularization_factors, val_score: 0.651254: 60%|###### | 12/20 [00:15<00:10, 1.25s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000381 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657638 valid's binary_logloss: 0.663101 [200] train's binary_logloss: 0.647154 valid's binary_logloss: 0.656096 [300] train's binary_logloss: 0.642213 valid's binary_logloss: 0.653702 [400] train's binary_logloss: 0.6386 valid's binary_logloss: 0.652862 [500] train's binary_logloss: 0.635493 valid's binary_logloss: 0.652354 [600] train's binary_logloss: 0.632518 valid's binary_logloss: 0.652123 [700] train's binary_logloss: 0.629817 valid's binary_logloss: 0.651851 [800] train's binary_logloss: 0.627236 valid's binary_logloss: 0.651695 Early stopping, best iteration is: [743] train's binary_logloss: 0.628666 valid's binary_logloss: 0.651254 regularization_factors, val_score: 0.651254: 65%|######5 | 13/20 [00:17<00:09, 1.34s/it][I 2020-09-27 04:40:02,095] Trial 52 finished with value: 0.651253816244402 and parameters: {'lambda_l1': 1.1047734679593184e-08, 'lambda_l2': 0.014255881410398455}. Best is trial 52 with value: 0.651253816244402. regularization_factors, val_score: 0.651254: 65%|######5 | 13/20 [00:17<00:09, 1.34s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000248 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657638 valid's binary_logloss: 0.663102 [200] train's binary_logloss: 0.647156 valid's binary_logloss: 0.656097 [300] train's binary_logloss: 0.642216 valid's binary_logloss: 0.653703 [400] train's binary_logloss: 0.638607 valid's binary_logloss: 0.652863 [500] train's binary_logloss: 0.6355 valid's binary_logloss: 0.652355 [600] train's binary_logloss: 0.632662 valid's binary_logloss: 0.652694 Early stopping, best iteration is: [531] train's binary_logloss: 0.634567 valid's binary_logloss: 0.652228 regularization_factors, val_score: 0.651254: 70%|####### | 14/20 [00:19<00:09, 1.53s/it][I 2020-09-27 04:40:04,065] Trial 53 finished with value: 0.6522275662505942 and parameters: {'lambda_l1': 1.4618448056501589e-08, 'lambda_l2': 0.04375470564403219}. Best is trial 52 with value: 0.651253816244402. regularization_factors, val_score: 0.651254: 70%|####### | 14/20 [00:19<00:09, 1.53s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000394 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657638 valid's binary_logloss: 0.663101 [200] train's binary_logloss: 0.647154 valid's binary_logloss: 0.656096 [300] train's binary_logloss: 0.642213 valid's binary_logloss: 0.653702 [400] train's binary_logloss: 0.6386 valid's binary_logloss: 0.652862 [500] train's binary_logloss: 0.635493 valid's binary_logloss: 0.652354 [600] train's binary_logloss: 0.632518 valid's binary_logloss: 0.652123 [700] train's binary_logloss: 0.629817 valid's binary_logloss: 0.651851 [800] train's binary_logloss: 0.627236 valid's binary_logloss: 0.651695 Early stopping, best iteration is: [743] train's binary_logloss: 0.628666 valid's binary_logloss: 0.651254 regularization_factors, val_score: 0.651254: 75%|#######5 | 15/20 [00:20<00:07, 1.45s/it][I 2020-09-27 04:40:05,343] Trial 54 finished with value: 0.6512538204787435 and parameters: {'lambda_l1': 1.0600410018512844e-08, 'lambda_l2': 0.014180637901064958}. Best is trial 52 with value: 0.651253816244402. regularization_factors, val_score: 0.651254: 75%|#######5 | 15/20 [00:20<00:07, 1.45s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000400 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657638 valid's binary_logloss: 0.663101 [200] train's binary_logloss: 0.647155 valid's binary_logloss: 0.656096 [300] train's binary_logloss: 0.642214 valid's binary_logloss: 0.653702 [400] train's binary_logloss: 0.638602 valid's binary_logloss: 0.652863 [500] train's binary_logloss: 0.635496 valid's binary_logloss: 0.652353 [600] train's binary_logloss: 0.632522 valid's binary_logloss: 0.652122 [700] train's binary_logloss: 0.629822 valid's binary_logloss: 0.65185 [800] train's binary_logloss: 0.627242 valid's binary_logloss: 0.651695 Early stopping, best iteration is: [743] train's binary_logloss: 0.628671 valid's binary_logloss: 0.651253 regularization_factors, val_score: 0.651253: 80%|######## | 16/20 [00:22<00:06, 1.51s/it][I 2020-09-27 04:40:06,971] Trial 55 finished with value: 0.6512532800647542 and parameters: {'lambda_l1': 1.1653169345025846e-08, 'lambda_l2': 0.02381661498148718}. Best is trial 55 with value: 0.6512532800647542. regularization_factors, val_score: 0.651253: 80%|######## | 16/20 [00:22<00:06, 1.51s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.014470 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657638 valid's binary_logloss: 0.663101 [200] train's binary_logloss: 0.647155 valid's binary_logloss: 0.656096 [300] train's binary_logloss: 0.642213 valid's binary_logloss: 0.653702 [400] train's binary_logloss: 0.638601 valid's binary_logloss: 0.652863 [500] train's binary_logloss: 0.635495 valid's binary_logloss: 0.652353 [600] train's binary_logloss: 0.632521 valid's binary_logloss: 0.652122 [700] train's binary_logloss: 0.62982 valid's binary_logloss: 0.65185 [800] train's binary_logloss: 0.62724 valid's binary_logloss: 0.651695 Early stopping, best iteration is: [743] train's binary_logloss: 0.628669 valid's binary_logloss: 0.651253 regularization_factors, val_score: 0.651253: 85%|########5 | 17/20 [00:23<00:04, 1.46s/it][I 2020-09-27 04:40:08,339] Trial 56 finished with value: 0.6512534743967614 and parameters: {'lambda_l1': 1.0004229968657813e-08, 'lambda_l2': 0.020345235153092046}. Best is trial 55 with value: 0.6512532800647542. regularization_factors, val_score: 0.651253: 85%|########5 | 17/20 [00:23<00:04, 1.46s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000378 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657638 valid's binary_logloss: 0.663101 [200] train's binary_logloss: 0.647155 valid's binary_logloss: 0.656096 [300] train's binary_logloss: 0.642213 valid's binary_logloss: 0.653702 [400] train's binary_logloss: 0.638602 valid's binary_logloss: 0.652863 [500] train's binary_logloss: 0.635495 valid's binary_logloss: 0.652353 [600] train's binary_logloss: 0.632521 valid's binary_logloss: 0.652122 [700] train's binary_logloss: 0.629821 valid's binary_logloss: 0.65185 [800] train's binary_logloss: 0.62724 valid's binary_logloss: 0.651695 Early stopping, best iteration is: [743] train's binary_logloss: 0.62867 valid's binary_logloss: 0.651253 regularization_factors, val_score: 0.651253: 90%|######### | 18/20 [00:25<00:02, 1.42s/it][I 2020-09-27 04:40:09,647] Trial 57 finished with value: 0.6512534186191336 and parameters: {'lambda_l1': 1.573306324418443e-08, 'lambda_l2': 0.021340714973965415}. Best is trial 55 with value: 0.6512532800647542. regularization_factors, val_score: 0.651253: 90%|######### | 18/20 [00:25<00:02, 1.42s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004407 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657638 valid's binary_logloss: 0.663102 [200] train's binary_logloss: 0.647156 valid's binary_logloss: 0.656097 [300] train's binary_logloss: 0.642216 valid's binary_logloss: 0.653703 [400] train's binary_logloss: 0.638606 valid's binary_logloss: 0.652863 [500] train's binary_logloss: 0.635499 valid's binary_logloss: 0.652355 [600] train's binary_logloss: 0.632662 valid's binary_logloss: 0.652694 Early stopping, best iteration is: [531] train's binary_logloss: 0.634567 valid's binary_logloss: 0.652228 regularization_factors, val_score: 0.651253: 95%|#########5| 19/20 [00:26<00:01, 1.41s/it][I 2020-09-27 04:40:11,033] Trial 58 finished with value: 0.6522275964597894 and parameters: {'lambda_l1': 1.1442661122184645e-08, 'lambda_l2': 0.04237276613246146}. Best is trial 55 with value: 0.6512532800647542. regularization_factors, val_score: 0.651253: 95%|#########5| 19/20 [00:26<00:01, 1.41s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002336 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657713 valid's binary_logloss: 0.663154 [200] train's binary_logloss: 0.647209 valid's binary_logloss: 0.655905 [300] train's binary_logloss: 0.64221 valid's binary_logloss: 0.65381 [400] train's binary_logloss: 0.638641 valid's binary_logloss: 0.653241 Early stopping, best iteration is: [351] train's binary_logloss: 0.640312 valid's binary_logloss: 0.65316 regularization_factors, val_score: 0.651253: 100%|##########| 20/20 [00:27<00:00, 1.24s/it][I 2020-09-27 04:40:11,871] Trial 59 finished with value: 0.6531604651549241 and parameters: {'lambda_l1': 0.7583136516531738, 'lambda_l2': 0.017657519269218453}. Best is trial 55 with value: 0.6512532800647542. regularization_factors, val_score: 0.651253: 100%|##########| 20/20 [00:27<00:00, 1.36s/it] min_data_in_leaf, val_score: 0.651253: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000390 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657638 valid's binary_logloss: 0.663101 [200] train's binary_logloss: 0.647129 valid's binary_logloss: 0.656026 [300] train's binary_logloss: 0.642165 valid's binary_logloss: 0.653612 [400] train's binary_logloss: 0.638561 valid's binary_logloss: 0.653007 [500] train's binary_logloss: 0.635388 valid's binary_logloss: 0.652537 [600] train's binary_logloss: 0.63262 valid's binary_logloss: 0.652704 Early stopping, best iteration is: [575] train's binary_logloss: 0.633278 valid's binary_logloss: 0.652405 min_data_in_leaf, val_score: 0.651253: 20%|## | 1/5 [00:01<00:04, 1.04s/it][I 2020-09-27 04:40:12,924] Trial 60 finished with value: 0.6524054109680009 and parameters: {'min_child_samples': 5}. Best is trial 60 with value: 0.6524054109680009. min_data_in_leaf, val_score: 0.651253: 20%|## | 1/5 [00:01<00:04, 1.04s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000685 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657638 valid's binary_logloss: 0.663101 [200] train's binary_logloss: 0.647173 valid's binary_logloss: 0.656292 [300] train's binary_logloss: 0.64235 valid's binary_logloss: 0.653435 [400] train's binary_logloss: 0.639047 valid's binary_logloss: 0.652707 [500] train's binary_logloss: 0.636172 valid's binary_logloss: 0.651812 [600] train's binary_logloss: 0.633542 valid's binary_logloss: 0.651717 [700] train's binary_logloss: 0.631145 valid's binary_logloss: 0.651603 [800] train's binary_logloss: 0.628853 valid's binary_logloss: 0.651423 Early stopping, best iteration is: [737] train's binary_logloss: 0.630317 valid's binary_logloss: 0.651187 min_data_in_leaf, val_score: 0.651187: 40%|#### | 2/5 [00:02<00:03, 1.09s/it][I 2020-09-27 04:40:14,129] Trial 61 finished with value: 0.6511874541615081 and parameters: {'min_child_samples': 100}. Best is trial 61 with value: 0.6511874541615081. min_data_in_leaf, val_score: 0.651187: 40%|#### | 2/5 [00:02<00:03, 1.09s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.006001 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657638 valid's binary_logloss: 0.663101 [200] train's binary_logloss: 0.647155 valid's binary_logloss: 0.656096 [300] train's binary_logloss: 0.64222 valid's binary_logloss: 0.653679 [400] train's binary_logloss: 0.638577 valid's binary_logloss: 0.653154 [500] train's binary_logloss: 0.635489 valid's binary_logloss: 0.652771 [600] train's binary_logloss: 0.632657 valid's binary_logloss: 0.652758 Early stopping, best iteration is: [531] train's binary_logloss: 0.634621 valid's binary_logloss: 0.65239 min_data_in_leaf, val_score: 0.651187: 60%|###### | 3/5 [00:03<00:02, 1.18s/it][I 2020-09-27 04:40:15,534] Trial 62 finished with value: 0.6523903635256941 and parameters: {'min_child_samples': 25}. Best is trial 61 with value: 0.6511874541615081. min_data_in_leaf, val_score: 0.651187: 60%|###### | 3/5 [00:03<00:02, 1.18s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000824 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657638 valid's binary_logloss: 0.663101 [200] train's binary_logloss: 0.647129 valid's binary_logloss: 0.656026 [300] train's binary_logloss: 0.642173 valid's binary_logloss: 0.653464 [400] train's binary_logloss: 0.638672 valid's binary_logloss: 0.652989 [500] train's binary_logloss: 0.635515 valid's binary_logloss: 0.652344 [600] train's binary_logloss: 0.632548 valid's binary_logloss: 0.652081 [700] train's binary_logloss: 0.629711 valid's binary_logloss: 0.651529 [800] train's binary_logloss: 0.627105 valid's binary_logloss: 0.651486 Early stopping, best iteration is: [749] train's binary_logloss: 0.628438 valid's binary_logloss: 0.651112 min_data_in_leaf, val_score: 0.651112: 80%|######## | 4/5 [00:04<00:01, 1.19s/it][I 2020-09-27 04:40:16,745] Trial 63 finished with value: 0.6511117328857268 and parameters: {'min_child_samples': 10}. Best is trial 63 with value: 0.6511117328857268. min_data_in_leaf, val_score: 0.651112: 80%|######## | 4/5 [00:04<00:01, 1.19s/it][LightGBM] [Info] Number of positive: 13151, number of negative: 12848 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000400 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4237 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505827 -> initscore=0.023310 [LightGBM] [Info] Start training from score 0.023310 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657638 valid's binary_logloss: 0.663101 [200] train's binary_logloss: 0.647155 valid's binary_logloss: 0.656096 [300] train's binary_logloss: 0.642357 valid's binary_logloss: 0.653658 [400] train's binary_logloss: 0.638938 valid's binary_logloss: 0.653056 [500] train's binary_logloss: 0.636126 valid's binary_logloss: 0.652589 [600] train's binary_logloss: 0.633479 valid's binary_logloss: 0.652767 [700] train's binary_logloss: 0.630836 valid's binary_logloss: 0.652316 Early stopping, best iteration is: [693] train's binary_logloss: 0.631059 valid's binary_logloss: 0.652118 min_data_in_leaf, val_score: 0.651112: 100%|##########| 5/5 [00:06<00:00, 1.19s/it][I 2020-09-27 04:40:17,925] Trial 64 finished with value: 0.6521175158063565 and parameters: {'min_child_samples': 50}. Best is trial 63 with value: 0.6511117328857268. min_data_in_leaf, val_score: 0.651112: 100%|##########| 5/5 [00:06<00:00, 1.21s/it] Fold : 2 [I 2020-09-27 04:40:18,017] A new study created in memory with name: no-name-75b2cbed-db22-45df-bfc2-635871fa6cfc feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.007450 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.578542 valid's binary_logloss: 0.656491 Early stopping, best iteration is: [79] train's binary_logloss: 0.59215 valid's binary_logloss: 0.655948 feature_fraction, val_score: 0.655948: 14%|#4 | 1/7 [00:01<00:06, 1.04s/it][I 2020-09-27 04:40:19,072] Trial 0 finished with value: 0.655947777974299 and parameters: {'feature_fraction': 0.5}. Best is trial 0 with value: 0.655947777974299. feature_fraction, val_score: 0.655948: 14%|#4 | 1/7 [00:01<00:06, 1.04s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.007324 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.572993 valid's binary_logloss: 0.65828 Early stopping, best iteration is: [79] train's binary_logloss: 0.587161 valid's binary_logloss: 0.656458 feature_fraction, val_score: 0.655948: 29%|##8 | 2/7 [00:01<00:04, 1.12it/s][I 2020-09-27 04:40:19,615] Trial 1 finished with value: 0.6564576518870222 and parameters: {'feature_fraction': 0.8}. Best is trial 0 with value: 0.655947777974299. feature_fraction, val_score: 0.655948: 29%|##8 | 2/7 [00:01<00:04, 1.12it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.006440 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.570115 valid's binary_logloss: 0.66021 Early stopping, best iteration is: [64] train's binary_logloss: 0.595593 valid's binary_logloss: 0.658255 feature_fraction, val_score: 0.655948: 43%|####2 | 3/7 [00:02<00:03, 1.29it/s][I 2020-09-27 04:40:20,131] Trial 2 finished with value: 0.6582553269377266 and parameters: {'feature_fraction': 1.0}. Best is trial 0 with value: 0.655947777974299. feature_fraction, val_score: 0.655948: 43%|####2 | 3/7 [00:02<00:03, 1.29it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010302 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.575364 valid's binary_logloss: 0.659046 Early stopping, best iteration is: [68] train's binary_logloss: 0.596133 valid's binary_logloss: 0.65791 feature_fraction, val_score: 0.655948: 57%|#####7 | 4/7 [00:02<00:02, 1.47it/s][I 2020-09-27 04:40:20,586] Trial 3 finished with value: 0.6579096104153804 and parameters: {'feature_fraction': 0.7}. Best is trial 0 with value: 0.655947777974299. feature_fraction, val_score: 0.655948: 57%|#####7 | 4/7 [00:02<00:02, 1.47it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000900 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.572299 valid's binary_logloss: 0.658723 Early stopping, best iteration is: [74] train's binary_logloss: 0.590022 valid's binary_logloss: 0.656515 feature_fraction, val_score: 0.655948: 71%|#######1 | 5/7 [00:03<00:01, 1.58it/s][I 2020-09-27 04:40:21,113] Trial 4 finished with value: 0.6565150373220784 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 0 with value: 0.655947777974299. feature_fraction, val_score: 0.655948: 71%|#######1 | 5/7 [00:03<00:01, 1.58it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009831 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.576537 valid's binary_logloss: 0.658655 Early stopping, best iteration is: [67] train's binary_logloss: 0.598042 valid's binary_logloss: 0.657232 feature_fraction, val_score: 0.655948: 86%|########5 | 6/7 [00:03<00:00, 1.74it/s][I 2020-09-27 04:40:21,552] Trial 5 finished with value: 0.6572319404658664 and parameters: {'feature_fraction': 0.6}. Best is trial 0 with value: 0.655947777974299. feature_fraction, val_score: 0.655948: 86%|########5 | 6/7 [00:03<00:00, 1.74it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000333 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.583115 valid's binary_logloss: 0.656735 Early stopping, best iteration is: [85] train's binary_logloss: 0.592606 valid's binary_logloss: 0.655167 feature_fraction, val_score: 0.655167: 100%|##########| 7/7 [00:04<00:00, 1.37it/s][I 2020-09-27 04:40:22,641] Trial 6 finished with value: 0.6551672639343036 and parameters: {'feature_fraction': 0.4}. Best is trial 6 with value: 0.6551672639343036. feature_fraction, val_score: 0.655167: 100%|##########| 7/7 [00:04<00:00, 1.52it/s] num_leaves, val_score: 0.655167: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004364 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.362 valid's binary_logloss: 0.666306 Early stopping, best iteration is: [62] train's binary_logloss: 0.44475 valid's binary_logloss: 0.66107 num_leaves, val_score: 0.655167: 5%|5 | 1/20 [00:00<00:17, 1.09it/s][I 2020-09-27 04:40:23,577] Trial 7 finished with value: 0.6610695796505417 and parameters: {'num_leaves': 186}. Best is trial 7 with value: 0.6610695796505417. num_leaves, val_score: 0.655167: 5%|5 | 1/20 [00:00<00:17, 1.09it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000378 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.509613 valid's binary_logloss: 0.655007 Early stopping, best iteration is: [63] train's binary_logloss: 0.555305 valid's binary_logloss: 0.653724 num_leaves, val_score: 0.653724: 10%|# | 2/20 [00:01<00:14, 1.22it/s][I 2020-09-27 04:40:24,159] Trial 8 finished with value: 0.6537238118183929 and parameters: {'num_leaves': 71}. Best is trial 8 with value: 0.6537238118183929. num_leaves, val_score: 0.653724: 10%|# | 2/20 [00:01<00:14, 1.22it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000384 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.486851 valid's binary_logloss: 0.659048 Early stopping, best iteration is: [64] train's binary_logloss: 0.536523 valid's binary_logloss: 0.655529 num_leaves, val_score: 0.653724: 15%|#5 | 3/20 [00:02<00:13, 1.29it/s][I 2020-09-27 04:40:24,823] Trial 9 finished with value: 0.6555293293116815 and parameters: {'num_leaves': 85}. Best is trial 8 with value: 0.6537238118183929. num_leaves, val_score: 0.653724: 15%|#5 | 3/20 [00:02<00:13, 1.29it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000459 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.668828 valid's binary_logloss: 0.671652 [200] train's binary_logloss: 0.659076 valid's binary_logloss: 0.663493 [300] train's binary_logloss: 0.653776 valid's binary_logloss: 0.659129 [400] train's binary_logloss: 0.65062 valid's binary_logloss: 0.656814 [500] train's binary_logloss: 0.648621 valid's binary_logloss: 0.655373 [600] train's binary_logloss: 0.647288 valid's binary_logloss: 0.654268 [700] train's binary_logloss: 0.646365 valid's binary_logloss: 0.653758 [800] train's binary_logloss: 0.645671 valid's binary_logloss: 0.653488 [900] train's binary_logloss: 0.645111 valid's binary_logloss: 0.653271 [1000] train's binary_logloss: 0.644638 valid's binary_logloss: 0.653148 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.644638 valid's binary_logloss: 0.653148 num_leaves, val_score: 0.653148: 20%|## | 4/20 [00:03<00:15, 1.02it/s][I 2020-09-27 04:40:26,293] Trial 10 finished with value: 0.6531478306778341 and parameters: {'num_leaves': 2}. Best is trial 10 with value: 0.6531478306778341. num_leaves, val_score: 0.653148: 20%|## | 4/20 [00:03<00:15, 1.02it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001421 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.647264 valid's binary_logloss: 0.65894 [200] train's binary_logloss: 0.635718 valid's binary_logloss: 0.655218 [300] train's binary_logloss: 0.627699 valid's binary_logloss: 0.654772 [400] train's binary_logloss: 0.620605 valid's binary_logloss: 0.65463 Early stopping, best iteration is: [337] train's binary_logloss: 0.625076 valid's binary_logloss: 0.654329 num_leaves, val_score: 0.653148: 25%|##5 | 5/20 [00:04<00:14, 1.05it/s][I 2020-09-27 04:40:27,188] Trial 11 finished with value: 0.6543286934013544 and parameters: {'num_leaves': 5}. Best is trial 10 with value: 0.6531478306778341. num_leaves, val_score: 0.653148: 25%|##5 | 5/20 [00:04<00:14, 1.05it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000570 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.631583 valid's binary_logloss: 0.655099 [200] train's binary_logloss: 0.611849 valid's binary_logloss: 0.653955 Early stopping, best iteration is: [189] train's binary_logloss: 0.613791 valid's binary_logloss: 0.653569 num_leaves, val_score: 0.653148: 30%|### | 6/20 [00:05<00:11, 1.22it/s][I 2020-09-27 04:40:27,696] Trial 12 finished with value: 0.6535687066231268 and parameters: {'num_leaves': 10}. Best is trial 10 with value: 0.6531478306778341. num_leaves, val_score: 0.653148: 30%|### | 6/20 [00:05<00:11, 1.22it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000425 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.634671 valid's binary_logloss: 0.655145 [200] train's binary_logloss: 0.616535 valid's binary_logloss: 0.655474 Early stopping, best iteration is: [160] train's binary_logloss: 0.62285 valid's binary_logloss: 0.654646 num_leaves, val_score: 0.653148: 35%|###5 | 7/20 [00:05<00:09, 1.39it/s][I 2020-09-27 04:40:28,179] Trial 13 finished with value: 0.6546457645585857 and parameters: {'num_leaves': 9}. Best is trial 10 with value: 0.6531478306778341. num_leaves, val_score: 0.653148: 35%|###5 | 7/20 [00:05<00:09, 1.39it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000412 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.668828 valid's binary_logloss: 0.671652 [200] train's binary_logloss: 0.659076 valid's binary_logloss: 0.663493 [300] train's binary_logloss: 0.653776 valid's binary_logloss: 0.659129 [400] train's binary_logloss: 0.65062 valid's binary_logloss: 0.656814 [500] train's binary_logloss: 0.648621 valid's binary_logloss: 0.655373 [600] train's binary_logloss: 0.647288 valid's binary_logloss: 0.654268 [700] train's binary_logloss: 0.646365 valid's binary_logloss: 0.653758 [800] train's binary_logloss: 0.645671 valid's binary_logloss: 0.653488 [900] train's binary_logloss: 0.645111 valid's binary_logloss: 0.653271 [1000] train's binary_logloss: 0.644638 valid's binary_logloss: 0.653148 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.644638 valid's binary_logloss: 0.653148 num_leaves, val_score: 0.653148: 40%|#### | 8/20 [00:06<00:10, 1.12it/s][I 2020-09-27 04:40:29,465] Trial 14 finished with value: 0.6531478306778341 and parameters: {'num_leaves': 2}. Best is trial 10 with value: 0.6531478306778341. num_leaves, val_score: 0.653148: 40%|#### | 8/20 [00:06<00:10, 1.12it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000471 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.306465 valid's binary_logloss: 0.670298 Early stopping, best iteration is: [44] train's binary_logloss: 0.45679 valid's binary_logloss: 0.663698 num_leaves, val_score: 0.653148: 45%|####5 | 9/20 [00:08<00:12, 1.15s/it][I 2020-09-27 04:40:31,226] Trial 15 finished with value: 0.663697988152403 and parameters: {'num_leaves': 248}. Best is trial 10 with value: 0.6531478306778341. num_leaves, val_score: 0.653148: 45%|####5 | 9/20 [00:08<00:12, 1.15s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004276 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.520782 valid's binary_logloss: 0.657921 Early stopping, best iteration is: [85] train's binary_logloss: 0.536614 valid's binary_logloss: 0.656043 num_leaves, val_score: 0.653148: 50%|##### | 10/20 [00:09<00:09, 1.03it/s][I 2020-09-27 04:40:31,778] Trial 16 finished with value: 0.6560430059842651 and parameters: {'num_leaves': 64}. Best is trial 10 with value: 0.6531478306778341. num_leaves, val_score: 0.653148: 50%|##### | 10/20 [00:09<00:09, 1.03it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000459 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.414427 valid's binary_logloss: 0.657771 Early stopping, best iteration is: [82] train's binary_logloss: 0.446068 valid's binary_logloss: 0.655491 num_leaves, val_score: 0.653148: 55%|#####5 | 11/20 [00:10<00:08, 1.03it/s][I 2020-09-27 04:40:32,746] Trial 17 finished with value: 0.6554905655085496 and parameters: {'num_leaves': 138}. Best is trial 10 with value: 0.6531478306778341. num_leaves, val_score: 0.653148: 55%|#####5 | 11/20 [00:10<00:08, 1.03it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000382 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.57465 valid's binary_logloss: 0.656774 [200] train's binary_logloss: 0.519325 valid's binary_logloss: 0.66119 Early stopping, best iteration is: [115] train's binary_logloss: 0.565412 valid's binary_logloss: 0.656102 num_leaves, val_score: 0.653148: 60%|###### | 12/20 [00:10<00:06, 1.20it/s][I 2020-09-27 04:40:33,269] Trial 18 finished with value: 0.6561023872985662 and parameters: {'num_leaves': 35}. Best is trial 10 with value: 0.6531478306778341. num_leaves, val_score: 0.653148: 60%|###### | 12/20 [00:10<00:06, 1.20it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000377 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.431599 valid's binary_logloss: 0.661998 Early stopping, best iteration is: [51] train's binary_logloss: 0.522437 valid's binary_logloss: 0.656248 num_leaves, val_score: 0.653148: 65%|######5 | 13/20 [00:11<00:06, 1.03it/s][I 2020-09-27 04:40:34,569] Trial 19 finished with value: 0.6562478660337796 and parameters: {'num_leaves': 124}. Best is trial 10 with value: 0.6531478306778341. num_leaves, val_score: 0.653148: 65%|######5 | 13/20 [00:11<00:06, 1.03it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000466 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.563816 valid's binary_logloss: 0.657169 Early stopping, best iteration is: [88] train's binary_logloss: 0.572908 valid's binary_logloss: 0.656464 num_leaves, val_score: 0.653148: 70%|####### | 14/20 [00:12<00:05, 1.20it/s][I 2020-09-27 04:40:35,073] Trial 20 finished with value: 0.6564643963450036 and parameters: {'num_leaves': 40}. Best is trial 10 with value: 0.6531478306778341. num_leaves, val_score: 0.653148: 70%|####### | 14/20 [00:12<00:05, 1.20it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000380 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.668828 valid's binary_logloss: 0.671652 [200] train's binary_logloss: 0.659076 valid's binary_logloss: 0.663493 [300] train's binary_logloss: 0.653776 valid's binary_logloss: 0.659129 [400] train's binary_logloss: 0.65062 valid's binary_logloss: 0.656814 [500] train's binary_logloss: 0.648621 valid's binary_logloss: 0.655373 [600] train's binary_logloss: 0.647288 valid's binary_logloss: 0.654268 [700] train's binary_logloss: 0.646365 valid's binary_logloss: 0.653758 [800] train's binary_logloss: 0.645671 valid's binary_logloss: 0.653488 [900] train's binary_logloss: 0.645111 valid's binary_logloss: 0.653271 [1000] train's binary_logloss: 0.644638 valid's binary_logloss: 0.653148 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.644638 valid's binary_logloss: 0.653148 num_leaves, val_score: 0.653148: 75%|#######5 | 15/20 [00:13<00:04, 1.06it/s][I 2020-09-27 04:40:36,259] Trial 21 finished with value: 0.6531478306778341 and parameters: {'num_leaves': 2}. Best is trial 10 with value: 0.6531478306778341. num_leaves, val_score: 0.653148: 75%|#######5 | 15/20 [00:13<00:04, 1.06it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010841 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.63738 valid's binary_logloss: 0.656329 [200] train's binary_logloss: 0.621173 valid's binary_logloss: 0.654961 [300] train's binary_logloss: 0.608196 valid's binary_logloss: 0.655376 Early stopping, best iteration is: [225] train's binary_logloss: 0.617766 valid's binary_logloss: 0.654241 num_leaves, val_score: 0.653148: 80%|######## | 16/20 [00:14<00:03, 1.14it/s][I 2020-09-27 04:40:36,999] Trial 22 finished with value: 0.654241444152379 and parameters: {'num_leaves': 8}. Best is trial 10 with value: 0.6531478306778341. num_leaves, val_score: 0.653148: 80%|######## | 16/20 [00:14<00:03, 1.14it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000584 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.576413 valid's binary_logloss: 0.656352 Early stopping, best iteration is: [85] train's binary_logloss: 0.58641 valid's binary_logloss: 0.655563 num_leaves, val_score: 0.653148: 85%|########5 | 17/20 [00:15<00:02, 1.02it/s][I 2020-09-27 04:40:38,198] Trial 23 finished with value: 0.6555629950352285 and parameters: {'num_leaves': 34}. Best is trial 10 with value: 0.6531478306778341. num_leaves, val_score: 0.653148: 85%|########5 | 17/20 [00:15<00:02, 1.02it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000677 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.647264 valid's binary_logloss: 0.65894 [200] train's binary_logloss: 0.635718 valid's binary_logloss: 0.655218 [300] train's binary_logloss: 0.627699 valid's binary_logloss: 0.654772 [400] train's binary_logloss: 0.620605 valid's binary_logloss: 0.65463 Early stopping, best iteration is: [337] train's binary_logloss: 0.625076 valid's binary_logloss: 0.654329 num_leaves, val_score: 0.653148: 90%|######### | 18/20 [00:16<00:01, 1.07it/s][I 2020-09-27 04:40:39,032] Trial 24 finished with value: 0.6543286934013544 and parameters: {'num_leaves': 5}. Best is trial 10 with value: 0.6531478306778341. num_leaves, val_score: 0.653148: 90%|######### | 18/20 [00:16<00:01, 1.07it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000412 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.557014 valid's binary_logloss: 0.658549 Early stopping, best iteration is: [46] train's binary_logloss: 0.606508 valid's binary_logloss: 0.656484 num_leaves, val_score: 0.653148: 95%|#########5| 19/20 [00:16<00:00, 1.23it/s][I 2020-09-27 04:40:39,562] Trial 25 finished with value: 0.6564836527501194 and parameters: {'num_leaves': 44}. Best is trial 10 with value: 0.6531478306778341. num_leaves, val_score: 0.653148: 95%|#########5| 19/20 [00:16<00:00, 1.23it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000389 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.455025 valid's binary_logloss: 0.660325 Early stopping, best iteration is: [39] train's binary_logloss: 0.562754 valid's binary_logloss: 0.655052 num_leaves, val_score: 0.653148: 100%|##########| 20/20 [00:17<00:00, 1.29it/s][I 2020-09-27 04:40:40,252] Trial 26 finished with value: 0.655051684902722 and parameters: {'num_leaves': 107}. Best is trial 10 with value: 0.6531478306778341. num_leaves, val_score: 0.653148: 100%|##########| 20/20 [00:17<00:00, 1.14it/s] bagging, val_score: 0.653148: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000406 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.668417 valid's binary_logloss: 0.671074 [200] train's binary_logloss: 0.658108 valid's binary_logloss: 0.662187 [300] train's binary_logloss: 0.652572 valid's binary_logloss: 0.657758 [400] train's binary_logloss: 0.649457 valid's binary_logloss: 0.655824 [500] train's binary_logloss: 0.64748 valid's binary_logloss: 0.654202 [600] train's binary_logloss: 0.646194 valid's binary_logloss: 0.653541 [700] train's binary_logloss: 0.645284 valid's binary_logloss: 0.65328 [800] train's binary_logloss: 0.644551 valid's binary_logloss: 0.653263 [900] train's binary_logloss: 0.64391 valid's binary_logloss: 0.653167 [1000] train's binary_logloss: 0.643382 valid's binary_logloss: 0.653181 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.643382 valid's binary_logloss: 0.653181 bagging, val_score: 0.653148: 10%|# | 1/10 [00:01<00:14, 1.60s/it][I 2020-09-27 04:40:41,874] Trial 27 finished with value: 0.6531811973594188 and parameters: {'bagging_fraction': 0.8306103567609042, 'bagging_freq': 5}. Best is trial 27 with value: 0.6531811973594188. bagging, val_score: 0.653148: 10%|# | 1/10 [00:01<00:14, 1.60s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000338 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.666533 valid's binary_logloss: 0.669291 [200] train's binary_logloss: 0.655913 valid's binary_logloss: 0.659959 [300] train's binary_logloss: 0.650521 valid's binary_logloss: 0.656005 [400] train's binary_logloss: 0.647915 valid's binary_logloss: 0.654692 [500] train's binary_logloss: 0.646471 valid's binary_logloss: 0.65411 [600] train's binary_logloss: 0.645402 valid's binary_logloss: 0.654374 Early stopping, best iteration is: [511] train's binary_logloss: 0.646347 valid's binary_logloss: 0.653867 bagging, val_score: 0.653148: 20%|## | 2/10 [00:02<00:11, 1.42s/it][I 2020-09-27 04:40:42,870] Trial 28 finished with value: 0.6538669216016616 and parameters: {'bagging_fraction': 0.4517567998193014, 'bagging_freq': 1}. Best is trial 27 with value: 0.6531811973594188. bagging, val_score: 0.653148: 20%|## | 2/10 [00:02<00:11, 1.42s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000347 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667071 valid's binary_logloss: 0.669326 [200] train's binary_logloss: 0.656588 valid's binary_logloss: 0.660429 [300] train's binary_logloss: 0.651211 valid's binary_logloss: 0.656184 [400] train's binary_logloss: 0.648478 valid's binary_logloss: 0.653762 [500] train's binary_logloss: 0.646843 valid's binary_logloss: 0.653211 [600] train's binary_logloss: 0.645633 valid's binary_logloss: 0.652424 Early stopping, best iteration is: [573] train's binary_logloss: 0.645934 valid's binary_logloss: 0.652252 bagging, val_score: 0.652252: 30%|### | 3/10 [00:03<00:08, 1.25s/it][I 2020-09-27 04:40:43,705] Trial 29 finished with value: 0.65225160633841 and parameters: {'bagging_fraction': 0.4737096463396763, 'bagging_freq': 7}. Best is trial 29 with value: 0.65225160633841. bagging, val_score: 0.652252: 30%|### | 3/10 [00:03<00:08, 1.25s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000502 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.666936 valid's binary_logloss: 0.668846 [200] train's binary_logloss: 0.656163 valid's binary_logloss: 0.659914 [300] train's binary_logloss: 0.650959 valid's binary_logloss: 0.655368 [400] train's binary_logloss: 0.648469 valid's binary_logloss: 0.653546 [500] train's binary_logloss: 0.646883 valid's binary_logloss: 0.653403 Early stopping, best iteration is: [438] train's binary_logloss: 0.647744 valid's binary_logloss: 0.652963 bagging, val_score: 0.652252: 40%|#### | 4/10 [00:04<00:06, 1.07s/it][I 2020-09-27 04:40:44,379] Trial 30 finished with value: 0.6529627183444107 and parameters: {'bagging_fraction': 0.4264803439292804, 'bagging_freq': 7}. Best is trial 29 with value: 0.65225160633841. bagging, val_score: 0.652252: 40%|#### | 4/10 [00:04<00:06, 1.07s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000349 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.666824 valid's binary_logloss: 0.668808 [200] train's binary_logloss: 0.656165 valid's binary_logloss: 0.65966 [300] train's binary_logloss: 0.650811 valid's binary_logloss: 0.654901 [400] train's binary_logloss: 0.648589 valid's binary_logloss: 0.652893 [500] train's binary_logloss: 0.647086 valid's binary_logloss: 0.653134 [600] train's binary_logloss: 0.645858 valid's binary_logloss: 0.652547 [700] train's binary_logloss: 0.644904 valid's binary_logloss: 0.65326 Early stopping, best iteration is: [623] train's binary_logloss: 0.64558 valid's binary_logloss: 0.652267 bagging, val_score: 0.652252: 50%|##### | 5/10 [00:05<00:05, 1.18s/it][I 2020-09-27 04:40:45,808] Trial 31 finished with value: 0.6522673041330311 and parameters: {'bagging_fraction': 0.41942184957156203, 'bagging_freq': 7}. Best is trial 29 with value: 0.65225160633841. bagging, val_score: 0.652252: 50%|##### | 5/10 [00:05<00:05, 1.18s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000470 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.666715 valid's binary_logloss: 0.669238 [200] train's binary_logloss: 0.656103 valid's binary_logloss: 0.660182 [300] train's binary_logloss: 0.650838 valid's binary_logloss: 0.655485 [400] train's binary_logloss: 0.648497 valid's binary_logloss: 0.653732 [500] train's binary_logloss: 0.646909 valid's binary_logloss: 0.653654 Early stopping, best iteration is: [470] train's binary_logloss: 0.6473 valid's binary_logloss: 0.652896 bagging, val_score: 0.652252: 60%|###### | 6/10 [00:06<00:04, 1.07s/it][I 2020-09-27 04:40:46,607] Trial 32 finished with value: 0.6528955403221239 and parameters: {'bagging_fraction': 0.41256437409812413, 'bagging_freq': 7}. Best is trial 29 with value: 0.65225160633841. bagging, val_score: 0.652252: 60%|###### | 6/10 [00:06<00:04, 1.07s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000341 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.666696 valid's binary_logloss: 0.669282 [200] train's binary_logloss: 0.656159 valid's binary_logloss: 0.660236 [300] train's binary_logloss: 0.65083 valid's binary_logloss: 0.655589 [400] train's binary_logloss: 0.648563 valid's binary_logloss: 0.653633 [500] train's binary_logloss: 0.64694 valid's binary_logloss: 0.653341 Early stopping, best iteration is: [470] train's binary_logloss: 0.647303 valid's binary_logloss: 0.652827 bagging, val_score: 0.652252: 70%|####### | 7/10 [00:06<00:02, 1.07it/s][I 2020-09-27 04:40:47,241] Trial 33 finished with value: 0.6528271831886884 and parameters: {'bagging_fraction': 0.40860834213805497, 'bagging_freq': 7}. Best is trial 29 with value: 0.65225160633841. bagging, val_score: 0.652252: 70%|####### | 7/10 [00:06<00:02, 1.07it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000446 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.666727 valid's binary_logloss: 0.669333 [200] train's binary_logloss: 0.656066 valid's binary_logloss: 0.660295 [300] train's binary_logloss: 0.650735 valid's binary_logloss: 0.655605 [400] train's binary_logloss: 0.648394 valid's binary_logloss: 0.653499 [500] train's binary_logloss: 0.646916 valid's binary_logloss: 0.653049 Early stopping, best iteration is: [470] train's binary_logloss: 0.647295 valid's binary_logloss: 0.652461 bagging, val_score: 0.652252: 80%|######## | 8/10 [00:07<00:01, 1.15it/s][I 2020-09-27 04:40:47,944] Trial 34 finished with value: 0.6524608085550598 and parameters: {'bagging_fraction': 0.40094624308690235, 'bagging_freq': 7}. Best is trial 29 with value: 0.65225160633841. bagging, val_score: 0.652252: 80%|######## | 8/10 [00:07<00:01, 1.15it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000704 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.666735 valid's binary_logloss: 0.669024 [200] train's binary_logloss: 0.656104 valid's binary_logloss: 0.660308 [300] train's binary_logloss: 0.650686 valid's binary_logloss: 0.655603 [400] train's binary_logloss: 0.648444 valid's binary_logloss: 0.653985 [500] train's binary_logloss: 0.646852 valid's binary_logloss: 0.653461 Early stopping, best iteration is: [470] train's binary_logloss: 0.647262 valid's binary_logloss: 0.652979 bagging, val_score: 0.652252: 90%|######### | 9/10 [00:08<00:00, 1.21it/s][I 2020-09-27 04:40:48,677] Trial 35 finished with value: 0.6529787611145468 and parameters: {'bagging_fraction': 0.40896873900368547, 'bagging_freq': 7}. Best is trial 29 with value: 0.65225160633841. bagging, val_score: 0.652252: 90%|######### | 9/10 [00:08<00:00, 1.21it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010501 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667557 valid's binary_logloss: 0.670365 [200] train's binary_logloss: 0.656945 valid's binary_logloss: 0.660765 [300] train's binary_logloss: 0.651474 valid's binary_logloss: 0.656544 [400] train's binary_logloss: 0.648605 valid's binary_logloss: 0.653988 [500] train's binary_logloss: 0.646848 valid's binary_logloss: 0.653526 [600] train's binary_logloss: 0.645737 valid's binary_logloss: 0.653035 Early stopping, best iteration is: [588] train's binary_logloss: 0.645878 valid's binary_logloss: 0.652666 bagging, val_score: 0.652252: 100%|##########| 10/10 [00:09<00:00, 1.03it/s][I 2020-09-27 04:40:49,991] Trial 36 finished with value: 0.65266565454767 and parameters: {'bagging_fraction': 0.5936827208546511, 'bagging_freq': 7}. Best is trial 29 with value: 0.65225160633841. bagging, val_score: 0.652252: 100%|##########| 10/10 [00:09<00:00, 1.03it/s] feature_fraction_stage2, val_score: 0.652252: 0%| | 0/3 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.004126 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667157 valid's binary_logloss: 0.669629 [200] train's binary_logloss: 0.65666 valid's binary_logloss: 0.660691 [300] train's binary_logloss: 0.651252 valid's binary_logloss: 0.656139 [400] train's binary_logloss: 0.648542 valid's binary_logloss: 0.653738 [500] train's binary_logloss: 0.646904 valid's binary_logloss: 0.653287 [600] train's binary_logloss: 0.645658 valid's binary_logloss: 0.652441 [700] train's binary_logloss: 0.644752 valid's binary_logloss: 0.652982 Early stopping, best iteration is: [623] train's binary_logloss: 0.645415 valid's binary_logloss: 0.652211 feature_fraction_stage2, val_score: 0.652211: 33%|###3 | 1/3 [00:00<00:01, 1.14it/s][I 2020-09-27 04:40:50,883] Trial 37 finished with value: 0.6522110872949437 and parameters: {'feature_fraction': 0.41600000000000004}. Best is trial 37 with value: 0.6522110872949437. feature_fraction_stage2, val_score: 0.652211: 33%|###3 | 1/3 [00:00<00:01, 1.14it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000546 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667214 valid's binary_logloss: 0.669614 [200] train's binary_logloss: 0.656661 valid's binary_logloss: 0.660306 [300] train's binary_logloss: 0.651258 valid's binary_logloss: 0.656152 [400] train's binary_logloss: 0.648532 valid's binary_logloss: 0.653868 [500] train's binary_logloss: 0.64686 valid's binary_logloss: 0.65353 [600] train's binary_logloss: 0.645694 valid's binary_logloss: 0.652608 [700] train's binary_logloss: 0.644833 valid's binary_logloss: 0.653223 Early stopping, best iteration is: [622] train's binary_logloss: 0.645442 valid's binary_logloss: 0.652368 feature_fraction_stage2, val_score: 0.652211: 67%|######6 | 2/3 [00:01<00:00, 1.14it/s][I 2020-09-27 04:40:51,765] Trial 38 finished with value: 0.6523684425922491 and parameters: {'feature_fraction': 0.44800000000000006}. Best is trial 37 with value: 0.6522110872949437. feature_fraction_stage2, val_score: 0.652211: 67%|######6 | 2/3 [00:01<00:00, 1.14it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000441 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667214 valid's binary_logloss: 0.669614 [200] train's binary_logloss: 0.656661 valid's binary_logloss: 0.660306 [300] train's binary_logloss: 0.651258 valid's binary_logloss: 0.656152 [400] train's binary_logloss: 0.648532 valid's binary_logloss: 0.653868 [500] train's binary_logloss: 0.64686 valid's binary_logloss: 0.65353 [600] train's binary_logloss: 0.645694 valid's binary_logloss: 0.652608 [700] train's binary_logloss: 0.644833 valid's binary_logloss: 0.653223 Early stopping, best iteration is: [622] train's binary_logloss: 0.645442 valid's binary_logloss: 0.652368 feature_fraction_stage2, val_score: 0.652211: 100%|##########| 3/3 [00:02<00:00, 1.13it/s][I 2020-09-27 04:40:52,662] Trial 39 finished with value: 0.6523684425922491 and parameters: {'feature_fraction': 0.48000000000000004}. Best is trial 37 with value: 0.6522110872949437. feature_fraction_stage2, val_score: 0.652211: 100%|##########| 3/3 [00:02<00:00, 1.12it/s] regularization_factors, val_score: 0.652211: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000446 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667157 valid's binary_logloss: 0.669629 [200] train's binary_logloss: 0.65666 valid's binary_logloss: 0.660691 [300] train's binary_logloss: 0.651252 valid's binary_logloss: 0.656139 [400] train's binary_logloss: 0.648542 valid's binary_logloss: 0.653738 [500] train's binary_logloss: 0.646905 valid's binary_logloss: 0.653287 [600] train's binary_logloss: 0.645659 valid's binary_logloss: 0.652441 [700] train's binary_logloss: 0.644752 valid's binary_logloss: 0.652981 Early stopping, best iteration is: [623] train's binary_logloss: 0.645415 valid's binary_logloss: 0.652211 regularization_factors, val_score: 0.652211: 5%|5 | 1/20 [00:01<00:25, 1.34s/it][I 2020-09-27 04:40:54,031] Trial 40 finished with value: 0.6522109977626771 and parameters: {'lambda_l1': 0.00014164147515972837, 'lambda_l2': 0.005481479902021038}. Best is trial 40 with value: 0.6522109977626771. regularization_factors, val_score: 0.652211: 5%|5 | 1/20 [00:01<00:25, 1.34s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004749 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667157 valid's binary_logloss: 0.669629 [200] train's binary_logloss: 0.656661 valid's binary_logloss: 0.660691 [300] train's binary_logloss: 0.651253 valid's binary_logloss: 0.656139 [400] train's binary_logloss: 0.648543 valid's binary_logloss: 0.653738 [500] train's binary_logloss: 0.646906 valid's binary_logloss: 0.653287 [600] train's binary_logloss: 0.64566 valid's binary_logloss: 0.652441 [700] train's binary_logloss: 0.644753 valid's binary_logloss: 0.652981 Early stopping, best iteration is: [623] train's binary_logloss: 0.645416 valid's binary_logloss: 0.652211 regularization_factors, val_score: 0.652211: 10%|# | 2/20 [00:02<00:21, 1.20s/it][I 2020-09-27 04:40:54,895] Trial 41 finished with value: 0.6522109708484218 and parameters: {'lambda_l1': 0.00010302212398099651, 'lambda_l2': 0.007995173172021716}. Best is trial 41 with value: 0.6522109708484218. regularization_factors, val_score: 0.652211: 10%|# | 2/20 [00:02<00:21, 1.20s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000406 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667157 valid's binary_logloss: 0.669629 [200] train's binary_logloss: 0.656661 valid's binary_logloss: 0.660691 [300] train's binary_logloss: 0.651252 valid's binary_logloss: 0.656139 [400] train's binary_logloss: 0.648542 valid's binary_logloss: 0.653738 [500] train's binary_logloss: 0.646905 valid's binary_logloss: 0.653287 [600] train's binary_logloss: 0.645659 valid's binary_logloss: 0.652441 [700] train's binary_logloss: 0.644752 valid's binary_logloss: 0.652981 Early stopping, best iteration is: [623] train's binary_logloss: 0.645415 valid's binary_logloss: 0.652211 regularization_factors, val_score: 0.652211: 15%|#5 | 3/20 [00:03<00:18, 1.10s/it][I 2020-09-27 04:40:55,766] Trial 42 finished with value: 0.6522109575004142 and parameters: {'lambda_l1': 5.703427719344543e-05, 'lambda_l2': 0.01133484756475216}. Best is trial 42 with value: 0.6522109575004142. regularization_factors, val_score: 0.652211: 15%|#5 | 3/20 [00:03<00:18, 1.10s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001530 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667157 valid's binary_logloss: 0.669629 [200] train's binary_logloss: 0.656661 valid's binary_logloss: 0.660691 [300] train's binary_logloss: 0.651252 valid's binary_logloss: 0.656139 [400] train's binary_logloss: 0.648542 valid's binary_logloss: 0.653738 [500] train's binary_logloss: 0.646905 valid's binary_logloss: 0.653287 [600] train's binary_logloss: 0.645659 valid's binary_logloss: 0.652441 [700] train's binary_logloss: 0.644753 valid's binary_logloss: 0.652981 Early stopping, best iteration is: [623] train's binary_logloss: 0.645416 valid's binary_logloss: 0.652211 regularization_factors, val_score: 0.652211: 20%|## | 4/20 [00:04<00:17, 1.08s/it][I 2020-09-27 04:40:56,793] Trial 43 finished with value: 0.6522109461124184 and parameters: {'lambda_l1': 7.037388835726531e-05, 'lambda_l2': 0.009023034240515858}. Best is trial 43 with value: 0.6522109461124184. regularization_factors, val_score: 0.652211: 20%|## | 4/20 [00:04<00:17, 1.08s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.017232 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667157 valid's binary_logloss: 0.669629 [200] train's binary_logloss: 0.656661 valid's binary_logloss: 0.660691 [300] train's binary_logloss: 0.651252 valid's binary_logloss: 0.656139 [400] train's binary_logloss: 0.648542 valid's binary_logloss: 0.653738 [500] train's binary_logloss: 0.646905 valid's binary_logloss: 0.653287 [600] train's binary_logloss: 0.645659 valid's binary_logloss: 0.652441 [700] train's binary_logloss: 0.644753 valid's binary_logloss: 0.652981 Early stopping, best iteration is: [623] train's binary_logloss: 0.645416 valid's binary_logloss: 0.652211 regularization_factors, val_score: 0.652211: 25%|##5 | 5/20 [00:05<00:16, 1.12s/it][I 2020-09-27 04:40:57,991] Trial 44 finished with value: 0.6522108862289743 and parameters: {'lambda_l1': 4.695264971426557e-05, 'lambda_l2': 0.012990997644376091}. Best is trial 44 with value: 0.6522108862289743. regularization_factors, val_score: 0.652211: 25%|##5 | 5/20 [00:05<00:16, 1.12s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000418 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667157 valid's binary_logloss: 0.669629 [200] train's binary_logloss: 0.656661 valid's binary_logloss: 0.660691 [300] train's binary_logloss: 0.651252 valid's binary_logloss: 0.656139 [400] train's binary_logloss: 0.648542 valid's binary_logloss: 0.653738 [500] train's binary_logloss: 0.646905 valid's binary_logloss: 0.653287 [600] train's binary_logloss: 0.645659 valid's binary_logloss: 0.652441 [700] train's binary_logloss: 0.644753 valid's binary_logloss: 0.652981 Early stopping, best iteration is: [623] train's binary_logloss: 0.645416 valid's binary_logloss: 0.652211 regularization_factors, val_score: 0.652211: 30%|### | 6/20 [00:06<00:14, 1.05s/it][I 2020-09-27 04:40:58,893] Trial 45 finished with value: 0.6522109405266442 and parameters: {'lambda_l1': 6.628550474869429e-05, 'lambda_l2': 0.00939838589902522}. Best is trial 44 with value: 0.6522108862289743. regularization_factors, val_score: 0.652211: 30%|### | 6/20 [00:06<00:14, 1.05s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000752 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667157 valid's binary_logloss: 0.669629 [200] train's binary_logloss: 0.656661 valid's binary_logloss: 0.660691 [300] train's binary_logloss: 0.651252 valid's binary_logloss: 0.656139 [400] train's binary_logloss: 0.648542 valid's binary_logloss: 0.653738 [500] train's binary_logloss: 0.646905 valid's binary_logloss: 0.653287 [600] train's binary_logloss: 0.645659 valid's binary_logloss: 0.652441 [700] train's binary_logloss: 0.644753 valid's binary_logloss: 0.652981 Early stopping, best iteration is: [623] train's binary_logloss: 0.645416 valid's binary_logloss: 0.652211 regularization_factors, val_score: 0.652211: 35%|###5 | 7/20 [00:07<00:13, 1.01s/it][I 2020-09-27 04:40:59,810] Trial 46 finished with value: 0.6522109149315195 and parameters: {'lambda_l1': 5.01427220376058e-05, 'lambda_l2': 0.011108962252046103}. Best is trial 44 with value: 0.6522108862289743. regularization_factors, val_score: 0.652211: 35%|###5 | 7/20 [00:07<00:13, 1.01s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000545 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667157 valid's binary_logloss: 0.66963 [200] train's binary_logloss: 0.656661 valid's binary_logloss: 0.660691 [300] train's binary_logloss: 0.651253 valid's binary_logloss: 0.656139 [400] train's binary_logloss: 0.648542 valid's binary_logloss: 0.653739 [500] train's binary_logloss: 0.646905 valid's binary_logloss: 0.653287 [600] train's binary_logloss: 0.64566 valid's binary_logloss: 0.652441 [700] train's binary_logloss: 0.644753 valid's binary_logloss: 0.652981 Early stopping, best iteration is: [623] train's binary_logloss: 0.645416 valid's binary_logloss: 0.652211 regularization_factors, val_score: 0.652211: 40%|#### | 8/20 [00:08<00:12, 1.03s/it][I 2020-09-27 04:41:00,882] Trial 47 finished with value: 0.6522108308126086 and parameters: {'lambda_l1': 4.076885797493761e-05, 'lambda_l2': 0.016633645127292034}. Best is trial 47 with value: 0.6522108308126086. regularization_factors, val_score: 0.652211: 40%|#### | 8/20 [00:08<00:12, 1.03s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000476 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667157 valid's binary_logloss: 0.66963 [200] train's binary_logloss: 0.656661 valid's binary_logloss: 0.660691 [300] train's binary_logloss: 0.651253 valid's binary_logloss: 0.656139 [400] train's binary_logloss: 0.648543 valid's binary_logloss: 0.653739 [500] train's binary_logloss: 0.646906 valid's binary_logloss: 0.653287 [600] train's binary_logloss: 0.64566 valid's binary_logloss: 0.652441 [700] train's binary_logloss: 0.644754 valid's binary_logloss: 0.652981 Early stopping, best iteration is: [623] train's binary_logloss: 0.645417 valid's binary_logloss: 0.652211 regularization_factors, val_score: 0.652211: 45%|####5 | 9/20 [00:09<00:11, 1.04s/it][I 2020-09-27 04:41:01,942] Trial 48 finished with value: 0.6522107665018237 and parameters: {'lambda_l1': 4.1633476557442786e-05, 'lambda_l2': 0.020110745479315996}. Best is trial 48 with value: 0.6522107665018237. regularization_factors, val_score: 0.652211: 45%|####5 | 9/20 [00:09<00:11, 1.04s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000514 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667174 valid's binary_logloss: 0.669606 [200] train's binary_logloss: 0.656682 valid's binary_logloss: 0.660677 [300] train's binary_logloss: 0.651269 valid's binary_logloss: 0.656126 [400] train's binary_logloss: 0.64857 valid's binary_logloss: 0.653596 [500] train's binary_logloss: 0.646941 valid's binary_logloss: 0.653105 [600] train's binary_logloss: 0.645643 valid's binary_logloss: 0.652442 [700] train's binary_logloss: 0.644768 valid's binary_logloss: 0.653208 Early stopping, best iteration is: [623] train's binary_logloss: 0.645403 valid's binary_logloss: 0.652216 regularization_factors, val_score: 0.652211: 50%|##### | 10/20 [00:10<00:09, 1.02it/s][I 2020-09-27 04:41:02,794] Trial 49 finished with value: 0.6522161063945356 and parameters: {'lambda_l1': 1.0048108252149604e-05, 'lambda_l2': 0.04595411901882225}. Best is trial 48 with value: 0.6522107665018237. regularization_factors, val_score: 0.652211: 50%|##### | 10/20 [00:10<00:09, 1.02it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000250 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667264 valid's binary_logloss: 0.669374 [200] train's binary_logloss: 0.656765 valid's binary_logloss: 0.660326 [300] train's binary_logloss: 0.651459 valid's binary_logloss: 0.656068 [400] train's binary_logloss: 0.648733 valid's binary_logloss: 0.65386 [500] train's binary_logloss: 0.647116 valid's binary_logloss: 0.653305 [600] train's binary_logloss: 0.645917 valid's binary_logloss: 0.652237 [700] train's binary_logloss: 0.645032 valid's binary_logloss: 0.652691 Early stopping, best iteration is: [622] train's binary_logloss: 0.645713 valid's binary_logloss: 0.652093 regularization_factors, val_score: 0.652093: 55%|#####5 | 11/20 [00:10<00:08, 1.06it/s][I 2020-09-27 04:41:03,645] Trial 50 finished with value: 0.6520932820946773 and parameters: {'lambda_l1': 5.634084468530463e-06, 'lambda_l2': 2.653747794681544}. Best is trial 50 with value: 0.6520932820946773. regularization_factors, val_score: 0.652093: 55%|#####5 | 11/20 [00:10<00:08, 1.06it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002392 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667312 valid's binary_logloss: 0.66942 [200] train's binary_logloss: 0.656827 valid's binary_logloss: 0.660345 [300] train's binary_logloss: 0.651466 valid's binary_logloss: 0.656086 [400] train's binary_logloss: 0.648842 valid's binary_logloss: 0.653957 [500] train's binary_logloss: 0.647208 valid's binary_logloss: 0.653403 [600] train's binary_logloss: 0.646029 valid's binary_logloss: 0.652507 [700] train's binary_logloss: 0.645184 valid's binary_logloss: 0.653114 Early stopping, best iteration is: [622] train's binary_logloss: 0.645818 valid's binary_logloss: 0.652395 regularization_factors, val_score: 0.652093: 60%|###### | 12/20 [00:12<00:08, 1.04s/it][I 2020-09-27 04:41:04,925] Trial 51 finished with value: 0.6523945935028854 and parameters: {'lambda_l1': 7.71584097888483e-06, 'lambda_l2': 4.4305490881206495}. Best is trial 50 with value: 0.6520932820946773. regularization_factors, val_score: 0.652093: 60%|###### | 12/20 [00:12<00:08, 1.04s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013219 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667193 valid's binary_logloss: 0.669624 [200] train's binary_logloss: 0.656716 valid's binary_logloss: 0.660578 [300] train's binary_logloss: 0.651263 valid's binary_logloss: 0.656044 [400] train's binary_logloss: 0.648561 valid's binary_logloss: 0.654042 [500] train's binary_logloss: 0.646891 valid's binary_logloss: 0.653546 [600] train's binary_logloss: 0.645722 valid's binary_logloss: 0.652543 [700] train's binary_logloss: 0.644815 valid's binary_logloss: 0.653225 Early stopping, best iteration is: [623] train's binary_logloss: 0.645497 valid's binary_logloss: 0.652335 regularization_factors, val_score: 0.652093: 65%|######5 | 13/20 [00:13<00:08, 1.16s/it][I 2020-09-27 04:41:06,349] Trial 52 finished with value: 0.6523351333047752 and parameters: {'lambda_l1': 7.422157693379738e-06, 'lambda_l2': 0.7332370123602728}. Best is trial 50 with value: 0.6520932820946773. regularization_factors, val_score: 0.652093: 65%|######5 | 13/20 [00:13<00:08, 1.16s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000421 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667345 valid's binary_logloss: 0.66941 [200] train's binary_logloss: 0.656873 valid's binary_logloss: 0.660467 [300] train's binary_logloss: 0.651513 valid's binary_logloss: 0.655838 [400] train's binary_logloss: 0.648818 valid's binary_logloss: 0.653846 [500] train's binary_logloss: 0.647188 valid's binary_logloss: 0.653186 [600] train's binary_logloss: 0.646098 valid's binary_logloss: 0.652539 Early stopping, best iteration is: [573] train's binary_logloss: 0.646377 valid's binary_logloss: 0.652312 regularization_factors, val_score: 0.652093: 70%|####### | 14/20 [00:14<00:06, 1.07s/it][I 2020-09-27 04:41:07,202] Trial 53 finished with value: 0.6523123313741976 and parameters: {'lambda_l1': 1.6736112176387592, 'lambda_l2': 3.1184950223667215e-07}. Best is trial 50 with value: 0.6520932820946773. regularization_factors, val_score: 0.652093: 70%|####### | 14/20 [00:14<00:06, 1.07s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002056 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667157 valid's binary_logloss: 0.669629 [200] train's binary_logloss: 0.65666 valid's binary_logloss: 0.660691 [300] train's binary_logloss: 0.651252 valid's binary_logloss: 0.656139 [400] train's binary_logloss: 0.648542 valid's binary_logloss: 0.653738 [500] train's binary_logloss: 0.646905 valid's binary_logloss: 0.653287 [600] train's binary_logloss: 0.645659 valid's binary_logloss: 0.652441 [700] train's binary_logloss: 0.644752 valid's binary_logloss: 0.652981 Early stopping, best iteration is: [623] train's binary_logloss: 0.645415 valid's binary_logloss: 0.652211 regularization_factors, val_score: 0.652093: 75%|#######5 | 15/20 [00:15<00:05, 1.02s/it][I 2020-09-27 04:41:08,102] Trial 54 finished with value: 0.6522110253635155 and parameters: {'lambda_l1': 0.0016228255112486223, 'lambda_l2': 8.007222095393149e-05}. Best is trial 50 with value: 0.6520932820946773. regularization_factors, val_score: 0.652093: 75%|#######5 | 15/20 [00:15<00:05, 1.02s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000382 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667157 valid's binary_logloss: 0.669629 [200] train's binary_logloss: 0.65666 valid's binary_logloss: 0.660691 [300] train's binary_logloss: 0.651252 valid's binary_logloss: 0.656139 [400] train's binary_logloss: 0.648542 valid's binary_logloss: 0.653738 [500] train's binary_logloss: 0.646904 valid's binary_logloss: 0.653287 [600] train's binary_logloss: 0.645658 valid's binary_logloss: 0.652441 [700] train's binary_logloss: 0.644752 valid's binary_logloss: 0.652982 Early stopping, best iteration is: [623] train's binary_logloss: 0.645415 valid's binary_logloss: 0.652211 regularization_factors, val_score: 0.652093: 80%|######## | 16/20 [00:16<00:04, 1.09s/it][I 2020-09-27 04:41:09,376] Trial 55 finished with value: 0.6522110831327725 and parameters: {'lambda_l1': 8.98931163949704e-08, 'lambda_l2': 0.0002694496140730187}. Best is trial 50 with value: 0.6520932820946773. regularization_factors, val_score: 0.652093: 80%|######## | 16/20 [00:16<00:04, 1.09s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000250 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.66718 valid's binary_logloss: 0.669612 [200] train's binary_logloss: 0.656691 valid's binary_logloss: 0.660683 [300] train's binary_logloss: 0.651278 valid's binary_logloss: 0.656131 [400] train's binary_logloss: 0.648583 valid's binary_logloss: 0.653603 [500] train's binary_logloss: 0.646957 valid's binary_logloss: 0.653104 [600] train's binary_logloss: 0.645691 valid's binary_logloss: 0.652458 [700] train's binary_logloss: 0.644799 valid's binary_logloss: 0.653165 Early stopping, best iteration is: [623] train's binary_logloss: 0.645445 valid's binary_logloss: 0.65223 regularization_factors, val_score: 0.652093: 85%|########5 | 17/20 [00:17<00:03, 1.03s/it][I 2020-09-27 04:41:10,268] Trial 56 finished with value: 0.6522296062194786 and parameters: {'lambda_l1': 1.272264047860099e-06, 'lambda_l2': 0.29472834124394065}. Best is trial 50 with value: 0.6520932820946773. regularization_factors, val_score: 0.652093: 85%|########5 | 17/20 [00:17<00:03, 1.03s/it][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000520 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667157 valid's binary_logloss: 0.669629 [200] train's binary_logloss: 0.656661 valid's binary_logloss: 0.660691 [300] train's binary_logloss: 0.651252 valid's binary_logloss: 0.656139 [400] train's binary_logloss: 0.648542 valid's binary_logloss: 0.653738 [500] train's binary_logloss: 0.646905 valid's binary_logloss: 0.653287 [600] train's binary_logloss: 0.645659 valid's binary_logloss: 0.652441 [700] train's binary_logloss: 0.644752 valid's binary_logloss: 0.652981 Early stopping, best iteration is: [623] train's binary_logloss: 0.645415 valid's binary_logloss: 0.652211 regularization_factors, val_score: 0.652093: 90%|######### | 18/20 [00:18<00:01, 1.02it/s][I 2020-09-27 04:41:11,137] Trial 57 finished with value: 0.6522109880071963 and parameters: {'lambda_l1': 0.0023192016556508074, 'lambda_l2': 0.0008178552198766434}. Best is trial 50 with value: 0.6520932820946773. regularization_factors, val_score: 0.652093: 90%|######### | 18/20 [00:18<00:01, 1.02it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000580 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667176 valid's binary_logloss: 0.669608 [200] train's binary_logloss: 0.656684 valid's binary_logloss: 0.660679 [300] train's binary_logloss: 0.651272 valid's binary_logloss: 0.656127 [400] train's binary_logloss: 0.648574 valid's binary_logloss: 0.653598 [500] train's binary_logloss: 0.646946 valid's binary_logloss: 0.653105 [600] train's binary_logloss: 0.645649 valid's binary_logloss: 0.652439 [700] train's binary_logloss: 0.644774 valid's binary_logloss: 0.653205 Early stopping, best iteration is: [623] train's binary_logloss: 0.64541 valid's binary_logloss: 0.652214 regularization_factors, val_score: 0.652093: 95%|#########5| 19/20 [00:19<00:00, 1.05it/s][I 2020-09-27 04:41:12,010] Trial 58 finished with value: 0.6522142117867799 and parameters: {'lambda_l1': 0.0007602407297528348, 'lambda_l2': 0.11500147374550874}. Best is trial 50 with value: 0.6520932820946773. regularization_factors, val_score: 0.652093: 95%|#########5| 19/20 [00:19<00:00, 1.05it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005008 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667156 valid's binary_logloss: 0.669629 [200] train's binary_logloss: 0.656659 valid's binary_logloss: 0.66069 [300] train's binary_logloss: 0.651251 valid's binary_logloss: 0.656139 [400] train's binary_logloss: 0.64854 valid's binary_logloss: 0.653738 [500] train's binary_logloss: 0.646903 valid's binary_logloss: 0.653287 [600] train's binary_logloss: 0.645657 valid's binary_logloss: 0.652441 [700] train's binary_logloss: 0.64475 valid's binary_logloss: 0.652981 Early stopping, best iteration is: [623] train's binary_logloss: 0.645413 valid's binary_logloss: 0.652211 regularization_factors, val_score: 0.652093: 100%|##########| 20/20 [00:20<00:00, 1.07s/it][I 2020-09-27 04:41:13,360] Trial 59 finished with value: 0.6522109875781879 and parameters: {'lambda_l1': 1.3713971348787328e-06, 'lambda_l2': 0.0013479533247432005}. Best is trial 50 with value: 0.6520932820946773. regularization_factors, val_score: 0.652093: 100%|##########| 20/20 [00:20<00:00, 1.03s/it] min_data_in_leaf, val_score: 0.652093: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004509 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667264 valid's binary_logloss: 0.669374 [200] train's binary_logloss: 0.656765 valid's binary_logloss: 0.660326 [300] train's binary_logloss: 0.651459 valid's binary_logloss: 0.656068 [400] train's binary_logloss: 0.648733 valid's binary_logloss: 0.653788 [500] train's binary_logloss: 0.647111 valid's binary_logloss: 0.653197 [600] train's binary_logloss: 0.645996 valid's binary_logloss: 0.652398 Early stopping, best iteration is: [589] train's binary_logloss: 0.646096 valid's binary_logloss: 0.652267 min_data_in_leaf, val_score: 0.652093: 20%|## | 1/5 [00:00<00:03, 1.23it/s][I 2020-09-27 04:41:14,186] Trial 60 finished with value: 0.6522671792511908 and parameters: {'min_child_samples': 50}. Best is trial 60 with value: 0.6522671792511908. min_data_in_leaf, val_score: 0.652093: 20%|## | 1/5 [00:00<00:03, 1.23it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000404 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667264 valid's binary_logloss: 0.669374 [200] train's binary_logloss: 0.656765 valid's binary_logloss: 0.660326 [300] train's binary_logloss: 0.651459 valid's binary_logloss: 0.656068 [400] train's binary_logloss: 0.648733 valid's binary_logloss: 0.653788 [500] train's binary_logloss: 0.647134 valid's binary_logloss: 0.653299 [600] train's binary_logloss: 0.645934 valid's binary_logloss: 0.652389 Early stopping, best iteration is: [573] train's binary_logloss: 0.646258 valid's binary_logloss: 0.652282 min_data_in_leaf, val_score: 0.652093: 40%|#### | 2/5 [00:01<00:02, 1.24it/s][I 2020-09-27 04:41:14,982] Trial 61 finished with value: 0.6522824409223347 and parameters: {'min_child_samples': 25}. Best is trial 60 with value: 0.6522671792511908. min_data_in_leaf, val_score: 0.652093: 40%|#### | 2/5 [00:01<00:02, 1.24it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001187 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667264 valid's binary_logloss: 0.669374 [200] train's binary_logloss: 0.656765 valid's binary_logloss: 0.660326 [300] train's binary_logloss: 0.651459 valid's binary_logloss: 0.656068 [400] train's binary_logloss: 0.648733 valid's binary_logloss: 0.65386 [500] train's binary_logloss: 0.647145 valid's binary_logloss: 0.653231 [600] train's binary_logloss: 0.645937 valid's binary_logloss: 0.652184 Early stopping, best iteration is: [573] train's binary_logloss: 0.646242 valid's binary_logloss: 0.652 min_data_in_leaf, val_score: 0.652000: 60%|###### | 3/5 [00:02<00:01, 1.25it/s][I 2020-09-27 04:41:15,775] Trial 62 finished with value: 0.6520002540523853 and parameters: {'min_child_samples': 10}. Best is trial 62 with value: 0.6520002540523853. min_data_in_leaf, val_score: 0.652000: 60%|###### | 3/5 [00:02<00:01, 1.25it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000591 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667264 valid's binary_logloss: 0.669374 [200] train's binary_logloss: 0.656765 valid's binary_logloss: 0.660326 [300] train's binary_logloss: 0.651468 valid's binary_logloss: 0.656076 [400] train's binary_logloss: 0.648844 valid's binary_logloss: 0.65385 [500] train's binary_logloss: 0.647295 valid's binary_logloss: 0.653583 [600] train's binary_logloss: 0.646236 valid's binary_logloss: 0.652591 Early stopping, best iteration is: [572] train's binary_logloss: 0.64651 valid's binary_logloss: 0.652484 min_data_in_leaf, val_score: 0.652000: 80%|######## | 4/5 [00:03<00:00, 1.07it/s][I 2020-09-27 04:41:17,010] Trial 63 finished with value: 0.6524841574858531 and parameters: {'min_child_samples': 100}. Best is trial 62 with value: 0.6520002540523853. min_data_in_leaf, val_score: 0.652000: 80%|######## | 4/5 [00:03<00:00, 1.07it/s][LightGBM] [Info] Number of positive: 13149, number of negative: 12850 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000499 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505750 -> initscore=0.023002 [LightGBM] [Info] Start training from score 0.023002 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667264 valid's binary_logloss: 0.669374 [200] train's binary_logloss: 0.656765 valid's binary_logloss: 0.660326 [300] train's binary_logloss: 0.651459 valid's binary_logloss: 0.656068 [400] train's binary_logloss: 0.648733 valid's binary_logloss: 0.65386 [500] train's binary_logloss: 0.647145 valid's binary_logloss: 0.653231 [600] train's binary_logloss: 0.645937 valid's binary_logloss: 0.652184 Early stopping, best iteration is: [573] train's binary_logloss: 0.646242 valid's binary_logloss: 0.652 min_data_in_leaf, val_score: 0.652000: 100%|##########| 5/5 [00:04<00:00, 1.13it/s][I 2020-09-27 04:41:17,796] Trial 64 finished with value: 0.6520002540523853 and parameters: {'min_child_samples': 5}. Best is trial 62 with value: 0.6520002540523853. min_data_in_leaf, val_score: 0.652000: 100%|##########| 5/5 [00:04<00:00, 1.13it/s] Fold : 3 [I 2020-09-27 04:41:17,860] A new study created in memory with name: no-name-2f38a2a7-25fd-480a-a0ca-a8d760f2fdde feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004894 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.572639 valid's binary_logloss: 0.657549 Early stopping, best iteration is: [69] train's binary_logloss: 0.593294 valid's binary_logloss: 0.65643 feature_fraction, val_score: 0.656430: 14%|#4 | 1/7 [00:00<00:03, 1.89it/s][I 2020-09-27 04:41:18,402] Trial 0 finished with value: 0.6564300006090903 and parameters: {'feature_fraction': 0.8}. Best is trial 0 with value: 0.6564300006090903. feature_fraction, val_score: 0.656430: 14%|#4 | 1/7 [00:00<00:03, 1.89it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000471 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.582961 valid's binary_logloss: 0.657331 [200] train's binary_logloss: 0.532881 valid's binary_logloss: 0.663509 Early stopping, best iteration is: [105] train's binary_logloss: 0.579867 valid's binary_logloss: 0.657057 feature_fraction, val_score: 0.656430: 29%|##8 | 2/7 [00:01<00:02, 1.88it/s][I 2020-09-27 04:41:18,934] Trial 1 finished with value: 0.6570566397890953 and parameters: {'feature_fraction': 0.4}. Best is trial 0 with value: 0.6564300006090903. feature_fraction, val_score: 0.656430: 29%|##8 | 2/7 [00:01<00:02, 1.88it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000466 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.578663 valid's binary_logloss: 0.658469 Early stopping, best iteration is: [94] train's binary_logloss: 0.582397 valid's binary_logloss: 0.657964 feature_fraction, val_score: 0.656430: 43%|####2 | 3/7 [00:01<00:02, 1.85it/s][I 2020-09-27 04:41:19,494] Trial 2 finished with value: 0.6579636929665533 and parameters: {'feature_fraction': 0.5}. Best is trial 0 with value: 0.6564300006090903. feature_fraction, val_score: 0.656430: 43%|####2 | 3/7 [00:01<00:02, 1.85it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.008636 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.571407 valid's binary_logloss: 0.662124 Early stopping, best iteration is: [73] train's binary_logloss: 0.589006 valid's binary_logloss: 0.660046 feature_fraction, val_score: 0.656430: 57%|#####7 | 4/7 [00:02<00:02, 1.47it/s][I 2020-09-27 04:41:20,502] Trial 3 finished with value: 0.6600459491612519 and parameters: {'feature_fraction': 1.0}. Best is trial 0 with value: 0.6564300006090903. feature_fraction, val_score: 0.656430: 57%|#####7 | 4/7 [00:02<00:02, 1.47it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.007591 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.576825 valid's binary_logloss: 0.659351 Early stopping, best iteration is: [59] train's binary_logloss: 0.60446 valid's binary_logloss: 0.656858 feature_fraction, val_score: 0.656430: 71%|#######1 | 5/7 [00:03<00:01, 1.57it/s][I 2020-09-27 04:41:21,037] Trial 4 finished with value: 0.6568575848518106 and parameters: {'feature_fraction': 0.6}. Best is trial 0 with value: 0.6564300006090903. feature_fraction, val_score: 0.656430: 71%|#######1 | 5/7 [00:03<00:01, 1.57it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.005521 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.574018 valid's binary_logloss: 0.658815 Early stopping, best iteration is: [58] train's binary_logloss: 0.603435 valid's binary_logloss: 0.65692 feature_fraction, val_score: 0.656430: 86%|########5 | 6/7 [00:03<00:00, 1.70it/s][I 2020-09-27 04:41:21,513] Trial 5 finished with value: 0.6569204528754933 and parameters: {'feature_fraction': 0.7}. Best is trial 0 with value: 0.6564300006090903. feature_fraction, val_score: 0.656430: 86%|########5 | 6/7 [00:03<00:00, 1.70it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000901 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.571583 valid's binary_logloss: 0.66348 Early stopping, best iteration is: [58] train's binary_logloss: 0.601331 valid's binary_logloss: 0.660277 feature_fraction, val_score: 0.656430: 100%|##########| 7/7 [00:04<00:00, 1.81it/s][I 2020-09-27 04:41:21,986] Trial 6 finished with value: 0.660277034185044 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 0 with value: 0.6564300006090903. feature_fraction, val_score: 0.656430: 100%|##########| 7/7 [00:04<00:00, 1.70it/s] num_leaves, val_score: 0.656430: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000435 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.316367 valid's binary_logloss: 0.678402 Early stopping, best iteration is: [38] train's binary_logloss: 0.477737 valid's binary_logloss: 0.661617 num_leaves, val_score: 0.656430: 5%|5 | 1/20 [00:01<00:22, 1.18s/it][I 2020-09-27 04:41:23,181] Trial 7 finished with value: 0.6616171717247548 and parameters: {'num_leaves': 199}. Best is trial 7 with value: 0.6616171717247548. num_leaves, val_score: 0.656430: 5%|5 | 1/20 [00:01<00:22, 1.18s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000849 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.294294 valid's binary_logloss: 0.683498 Early stopping, best iteration is: [30] train's binary_logloss: 0.496866 valid's binary_logloss: 0.663982 num_leaves, val_score: 0.656430: 10%|# | 2/20 [00:02<00:24, 1.34s/it][I 2020-09-27 04:41:24,901] Trial 8 finished with value: 0.6639821465726022 and parameters: {'num_leaves': 220}. Best is trial 7 with value: 0.6616171717247548. num_leaves, val_score: 0.656430: 10%|# | 2/20 [00:02<00:24, 1.34s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000432 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.317927 valid's binary_logloss: 0.67117 Early stopping, best iteration is: [38] train's binary_logloss: 0.478124 valid's binary_logloss: 0.658535 num_leaves, val_score: 0.656430: 15%|#5 | 3/20 [00:04<00:21, 1.28s/it][I 2020-09-27 04:41:26,051] Trial 9 finished with value: 0.6585349270193589 and parameters: {'num_leaves': 197}. Best is trial 9 with value: 0.6585349270193589. num_leaves, val_score: 0.656430: 15%|#5 | 3/20 [00:04<00:21, 1.28s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000872 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.6214 valid's binary_logloss: 0.657582 [200] train's binary_logloss: 0.595975 valid's binary_logloss: 0.659137 Early stopping, best iteration is: [112] train's binary_logloss: 0.618201 valid's binary_logloss: 0.656843 num_leaves, val_score: 0.656430: 20%|## | 4/20 [00:04<00:16, 1.04s/it][I 2020-09-27 04:41:26,504] Trial 10 finished with value: 0.6568431632806528 and parameters: {'num_leaves': 12}. Best is trial 10 with value: 0.6568431632806528. num_leaves, val_score: 0.656430: 20%|## | 4/20 [00:04<00:16, 1.04s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000825 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.633772 valid's binary_logloss: 0.658331 [200] train's binary_logloss: 0.615857 valid's binary_logloss: 0.657932 Early stopping, best iteration is: [151] train's binary_logloss: 0.623812 valid's binary_logloss: 0.657048 num_leaves, val_score: 0.656430: 25%|##5 | 5/20 [00:04<00:12, 1.16it/s][I 2020-09-27 04:41:26,971] Trial 11 finished with value: 0.6570482164914353 and parameters: {'num_leaves': 8}. Best is trial 10 with value: 0.6568431632806528. num_leaves, val_score: 0.656430: 25%|##5 | 5/20 [00:04<00:12, 1.16it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008860 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.6214 valid's binary_logloss: 0.657582 [200] train's binary_logloss: 0.595975 valid's binary_logloss: 0.659137 Early stopping, best iteration is: [112] train's binary_logloss: 0.618201 valid's binary_logloss: 0.656843 num_leaves, val_score: 0.656430: 30%|### | 6/20 [00:05<00:11, 1.18it/s][I 2020-09-27 04:41:27,770] Trial 12 finished with value: 0.6568431632806528 and parameters: {'num_leaves': 12}. Best is trial 10 with value: 0.6568431632806528. num_leaves, val_score: 0.656430: 30%|### | 6/20 [00:05<00:11, 1.18it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012578 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.458511 valid's binary_logloss: 0.667979 Early stopping, best iteration is: [42] train's binary_logloss: 0.55365 valid's binary_logloss: 0.659615 num_leaves, val_score: 0.656430: 35%|###5 | 7/20 [00:06<00:11, 1.14it/s][I 2020-09-27 04:41:28,720] Trial 13 finished with value: 0.6596145393331994 and parameters: {'num_leaves': 89}. Best is trial 10 with value: 0.6568431632806528. num_leaves, val_score: 0.656430: 35%|###5 | 7/20 [00:06<00:11, 1.14it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007578 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.458372 valid's binary_logloss: 0.664586 Early stopping, best iteration is: [37] train's binary_logloss: 0.562981 valid's binary_logloss: 0.659102 num_leaves, val_score: 0.656430: 40%|#### | 8/20 [00:07<00:09, 1.23it/s][I 2020-09-27 04:41:29,393] Trial 14 finished with value: 0.6591016497023785 and parameters: {'num_leaves': 90}. Best is trial 10 with value: 0.6568431632806528. num_leaves, val_score: 0.656430: 40%|#### | 8/20 [00:07<00:09, 1.23it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001036 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.500134 valid's binary_logloss: 0.660323 Early stopping, best iteration is: [55] train's binary_logloss: 0.556151 valid's binary_logloss: 0.658021 num_leaves, val_score: 0.656430: 45%|####5 | 9/20 [00:08<00:08, 1.29it/s][I 2020-09-27 04:41:30,070] Trial 15 finished with value: 0.6580206440647851 and parameters: {'num_leaves': 66}. Best is trial 10 with value: 0.6568431632806528. num_leaves, val_score: 0.656430: 45%|####5 | 9/20 [00:08<00:08, 1.29it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004995 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.37816 valid's binary_logloss: 0.67245 Early stopping, best iteration is: [33] train's binary_logloss: 0.531324 valid's binary_logloss: 0.658413 num_leaves, val_score: 0.656430: 50%|##### | 10/20 [00:08<00:07, 1.27it/s][I 2020-09-27 04:41:30,896] Trial 16 finished with value: 0.6584130655795914 and parameters: {'num_leaves': 144}. Best is trial 10 with value: 0.6568431632806528. num_leaves, val_score: 0.656430: 50%|##### | 10/20 [00:08<00:07, 1.27it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000912 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.557347 valid's binary_logloss: 0.658946 Early stopping, best iteration is: [52] train's binary_logloss: 0.597034 valid's binary_logloss: 0.657578 num_leaves, val_score: 0.656430: 55%|#####5 | 11/20 [00:09<00:07, 1.15it/s][I 2020-09-27 04:41:31,958] Trial 17 finished with value: 0.6575775290564063 and parameters: {'num_leaves': 38}. Best is trial 10 with value: 0.6568431632806528. num_leaves, val_score: 0.656430: 55%|#####5 | 11/20 [00:09<00:07, 1.15it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004745 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.380846 valid's binary_logloss: 0.673729 Early stopping, best iteration is: [29] train's binary_logloss: 0.545732 valid's binary_logloss: 0.661242 num_leaves, val_score: 0.656430: 60%|###### | 12/20 [00:10<00:06, 1.17it/s][I 2020-09-27 04:41:32,785] Trial 18 finished with value: 0.6612424428440397 and parameters: {'num_leaves': 143}. Best is trial 10 with value: 0.6568431632806528. num_leaves, val_score: 0.656430: 60%|###### | 12/20 [00:10<00:06, 1.17it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000446 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.270716 valid's binary_logloss: 0.684739 Early stopping, best iteration is: [22] train's binary_logloss: 0.522558 valid's binary_logloss: 0.66216 num_leaves, val_score: 0.656430: 65%|######5 | 13/20 [00:12<00:06, 1.03it/s][I 2020-09-27 04:41:34,008] Trial 19 finished with value: 0.6621600912726282 and parameters: {'num_leaves': 248}. Best is trial 10 with value: 0.6568431632806528. num_leaves, val_score: 0.656430: 65%|######5 | 13/20 [00:12<00:06, 1.03it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000999 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.53747 valid's binary_logloss: 0.659534 Early stopping, best iteration is: [53] train's binary_logloss: 0.583686 valid's binary_logloss: 0.65689 num_leaves, val_score: 0.656430: 70%|####### | 14/20 [00:12<00:05, 1.18it/s][I 2020-09-27 04:41:34,577] Trial 20 finished with value: 0.6568899584326288 and parameters: {'num_leaves': 47}. Best is trial 10 with value: 0.6568431632806528. num_leaves, val_score: 0.656430: 70%|####### | 14/20 [00:12<00:05, 1.18it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000858 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.668556 valid's binary_logloss: 0.673758 [200] train's binary_logloss: 0.658927 valid's binary_logloss: 0.665036 [300] train's binary_logloss: 0.653688 valid's binary_logloss: 0.660827 [400] train's binary_logloss: 0.650562 valid's binary_logloss: 0.658072 [500] train's binary_logloss: 0.648564 valid's binary_logloss: 0.656465 [600] train's binary_logloss: 0.647233 valid's binary_logloss: 0.655679 [700] train's binary_logloss: 0.646297 valid's binary_logloss: 0.655125 [800] train's binary_logloss: 0.64559 valid's binary_logloss: 0.654927 [900] train's binary_logloss: 0.645037 valid's binary_logloss: 0.654748 [1000] train's binary_logloss: 0.644582 valid's binary_logloss: 0.654744 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.644582 valid's binary_logloss: 0.654744 num_leaves, val_score: 0.654744: 75%|#######5 | 15/20 [00:14<00:05, 1.10s/it][I 2020-09-27 04:41:36,250] Trial 21 finished with value: 0.6547443275104798 and parameters: {'num_leaves': 2}. Best is trial 21 with value: 0.6547443275104798. num_leaves, val_score: 0.654744: 75%|#######5 | 15/20 [00:14<00:05, 1.10s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005030 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.607455 valid's binary_logloss: 0.658232 Early stopping, best iteration is: [89] train's binary_logloss: 0.611867 valid's binary_logloss: 0.657658 num_leaves, val_score: 0.654744: 80%|######## | 16/20 [00:14<00:03, 1.11it/s][I 2020-09-27 04:41:36,691] Trial 22 finished with value: 0.6576583665143456 and parameters: {'num_leaves': 17}. Best is trial 21 with value: 0.6547443275104798. num_leaves, val_score: 0.654744: 80%|######## | 16/20 [00:14<00:03, 1.11it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000984 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657455 valid's binary_logloss: 0.665049 [200] train's binary_logloss: 0.647037 valid's binary_logloss: 0.658049 [300] train's binary_logloss: 0.641503 valid's binary_logloss: 0.655542 [400] train's binary_logloss: 0.6376 valid's binary_logloss: 0.6548 [500] train's binary_logloss: 0.634488 valid's binary_logloss: 0.654825 [600] train's binary_logloss: 0.631676 valid's binary_logloss: 0.654591 Early stopping, best iteration is: [550] train's binary_logloss: 0.633043 valid's binary_logloss: 0.654226 num_leaves, val_score: 0.654226: 85%|########5 | 17/20 [00:15<00:02, 1.11it/s][I 2020-09-27 04:41:37,600] Trial 23 finished with value: 0.6542264931984162 and parameters: {'num_leaves': 3}. Best is trial 23 with value: 0.6542264931984162. num_leaves, val_score: 0.654226: 85%|########5 | 17/20 [00:15<00:02, 1.11it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000878 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.44017 valid's binary_logloss: 0.669582 Early stopping, best iteration is: [27] train's binary_logloss: 0.579264 valid's binary_logloss: 0.660699 num_leaves, val_score: 0.654226: 90%|######### | 18/20 [00:16<00:01, 1.17it/s][I 2020-09-27 04:41:38,351] Trial 24 finished with value: 0.6606988142412215 and parameters: {'num_leaves': 101}. Best is trial 23 with value: 0.6542264931984162. num_leaves, val_score: 0.654226: 90%|######### | 18/20 [00:16<00:01, 1.17it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000476 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.534907 valid's binary_logloss: 0.660247 Early stopping, best iteration is: [52] train's binary_logloss: 0.583748 valid's binary_logloss: 0.656588 num_leaves, val_score: 0.654226: 95%|#########5| 19/20 [00:17<00:00, 1.05it/s][I 2020-09-27 04:41:39,536] Trial 25 finished with value: 0.6565875567036938 and parameters: {'num_leaves': 48}. Best is trial 23 with value: 0.6542264931984162. num_leaves, val_score: 0.654226: 95%|#########5| 19/20 [00:17<00:00, 1.05it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000989 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.412133 valid's binary_logloss: 0.670637 Early stopping, best iteration is: [35] train's binary_logloss: 0.542676 valid's binary_logloss: 0.662727 num_leaves, val_score: 0.654226: 100%|##########| 20/20 [00:18<00:00, 1.09it/s][I 2020-09-27 04:41:40,357] Trial 26 finished with value: 0.6627267041691768 and parameters: {'num_leaves': 120}. Best is trial 23 with value: 0.6542264931984162. num_leaves, val_score: 0.654226: 100%|##########| 20/20 [00:18<00:00, 1.09it/s] bagging, val_score: 0.654226: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004771 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656611 valid's binary_logloss: 0.664545 [200] train's binary_logloss: 0.64587 valid's binary_logloss: 0.657646 [300] train's binary_logloss: 0.640645 valid's binary_logloss: 0.655721 [400] train's binary_logloss: 0.636853 valid's binary_logloss: 0.655725 [500] train's binary_logloss: 0.633496 valid's binary_logloss: 0.655542 [600] train's binary_logloss: 0.630016 valid's binary_logloss: 0.655279 Early stopping, best iteration is: [583] train's binary_logloss: 0.6307 valid's binary_logloss: 0.655229 bagging, val_score: 0.654226: 10%|# | 1/10 [00:00<00:08, 1.05it/s][I 2020-09-27 04:41:41,322] Trial 27 finished with value: 0.6552294862634005 and parameters: {'bagging_fraction': 0.7419038111841914, 'bagging_freq': 4}. Best is trial 27 with value: 0.6552294862634005. bagging, val_score: 0.654226: 10%|# | 1/10 [00:00<00:08, 1.05it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011398 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656644 valid's binary_logloss: 0.664211 [200] train's binary_logloss: 0.646379 valid's binary_logloss: 0.65746 [300] train's binary_logloss: 0.64107 valid's binary_logloss: 0.655347 [400] train's binary_logloss: 0.63725 valid's binary_logloss: 0.654521 [500] train's binary_logloss: 0.633892 valid's binary_logloss: 0.654747 Early stopping, best iteration is: [418] train's binary_logloss: 0.636564 valid's binary_logloss: 0.654327 bagging, val_score: 0.654226: 20%|## | 2/10 [00:01<00:07, 1.10it/s][I 2020-09-27 04:41:42,144] Trial 28 finished with value: 0.6543274423941562 and parameters: {'bagging_fraction': 0.7471333362321411, 'bagging_freq': 4}. Best is trial 28 with value: 0.6543274423941562. bagging, val_score: 0.654226: 20%|## | 2/10 [00:01<00:07, 1.10it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000868 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657402 valid's binary_logloss: 0.66564 [200] train's binary_logloss: 0.646902 valid's binary_logloss: 0.659056 [300] train's binary_logloss: 0.641447 valid's binary_logloss: 0.656325 [400] train's binary_logloss: 0.637573 valid's binary_logloss: 0.655624 [500] train's binary_logloss: 0.634342 valid's binary_logloss: 0.655314 Early stopping, best iteration is: [451] train's binary_logloss: 0.635853 valid's binary_logloss: 0.655055 bagging, val_score: 0.654226: 30%|### | 3/10 [00:03<00:07, 1.05s/it][I 2020-09-27 04:41:43,507] Trial 29 finished with value: 0.6550545130006391 and parameters: {'bagging_fraction': 0.9853823451672025, 'bagging_freq': 7}. Best is trial 28 with value: 0.6543274423941562. bagging, val_score: 0.654226: 30%|### | 3/10 [00:03<00:07, 1.05s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005058 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.655834 valid's binary_logloss: 0.664357 [200] train's binary_logloss: 0.645749 valid's binary_logloss: 0.656264 [300] train's binary_logloss: 0.641025 valid's binary_logloss: 0.656599 Early stopping, best iteration is: [244] train's binary_logloss: 0.643386 valid's binary_logloss: 0.656058 bagging, val_score: 0.654226: 40%|#### | 4/10 [00:03<00:05, 1.08it/s][I 2020-09-27 04:41:44,148] Trial 30 finished with value: 0.6560583873045698 and parameters: {'bagging_fraction': 0.5579781104550179, 'bagging_freq': 3}. Best is trial 28 with value: 0.6543274423941562. bagging, val_score: 0.654226: 40%|#### | 4/10 [00:03<00:05, 1.08it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000466 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657297 valid's binary_logloss: 0.665521 [200] train's binary_logloss: 0.64689 valid's binary_logloss: 0.658315 [300] train's binary_logloss: 0.641484 valid's binary_logloss: 0.655464 [400] train's binary_logloss: 0.637638 valid's binary_logloss: 0.654412 [500] train's binary_logloss: 0.634281 valid's binary_logloss: 0.653965 [600] train's binary_logloss: 0.631536 valid's binary_logloss: 0.653942 Early stopping, best iteration is: [511] train's binary_logloss: 0.633982 valid's binary_logloss: 0.65385 bagging, val_score: 0.653850: 50%|##### | 5/10 [00:04<00:04, 1.06it/s][I 2020-09-27 04:41:45,139] Trial 31 finished with value: 0.6538504332512516 and parameters: {'bagging_fraction': 0.9521342276139986, 'bagging_freq': 7}. Best is trial 31 with value: 0.6538504332512516. bagging, val_score: 0.653850: 50%|##### | 5/10 [00:04<00:04, 1.06it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005270 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657127 valid's binary_logloss: 0.665533 [200] train's binary_logloss: 0.646571 valid's binary_logloss: 0.65777 [300] train's binary_logloss: 0.641217 valid's binary_logloss: 0.655631 [400] train's binary_logloss: 0.637412 valid's binary_logloss: 0.654782 [500] train's binary_logloss: 0.63418 valid's binary_logloss: 0.654189 Early stopping, best iteration is: [483] train's binary_logloss: 0.634685 valid's binary_logloss: 0.654112 bagging, val_score: 0.653850: 60%|###### | 6/10 [00:05<00:03, 1.07it/s][I 2020-09-27 04:41:46,049] Trial 32 finished with value: 0.6541122278782455 and parameters: {'bagging_fraction': 0.9284471640035712, 'bagging_freq': 7}. Best is trial 31 with value: 0.6538504332512516. bagging, val_score: 0.653850: 60%|###### | 6/10 [00:05<00:03, 1.07it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000855 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657282 valid's binary_logloss: 0.665245 [200] train's binary_logloss: 0.646764 valid's binary_logloss: 0.657757 [300] train's binary_logloss: 0.641399 valid's binary_logloss: 0.655443 [400] train's binary_logloss: 0.637639 valid's binary_logloss: 0.654744 [500] train's binary_logloss: 0.634399 valid's binary_logloss: 0.654233 [600] train's binary_logloss: 0.63167 valid's binary_logloss: 0.654358 Early stopping, best iteration is: [506] train's binary_logloss: 0.634248 valid's binary_logloss: 0.654145 bagging, val_score: 0.653850: 70%|####### | 7/10 [00:07<00:03, 1.05s/it][I 2020-09-27 04:41:47,378] Trial 33 finished with value: 0.6541452383331995 and parameters: {'bagging_fraction': 0.9500310046900615, 'bagging_freq': 7}. Best is trial 31 with value: 0.6538504332512516. bagging, val_score: 0.653850: 70%|####### | 7/10 [00:07<00:03, 1.05s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004962 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657608 valid's binary_logloss: 0.665788 [200] train's binary_logloss: 0.647027 valid's binary_logloss: 0.6581 [300] train's binary_logloss: 0.641573 valid's binary_logloss: 0.655658 [400] train's binary_logloss: 0.637741 valid's binary_logloss: 0.655284 [500] train's binary_logloss: 0.634569 valid's binary_logloss: 0.654728 [600] train's binary_logloss: 0.631817 valid's binary_logloss: 0.654394 [700] train's binary_logloss: 0.629035 valid's binary_logloss: 0.654408 Early stopping, best iteration is: [653] train's binary_logloss: 0.630298 valid's binary_logloss: 0.654176 bagging, val_score: 0.653850: 80%|######## | 8/10 [00:08<00:02, 1.11s/it][I 2020-09-27 04:41:48,627] Trial 34 finished with value: 0.654175795227107 and parameters: {'bagging_fraction': 0.9963869875397509, 'bagging_freq': 7}. Best is trial 31 with value: 0.6538504332512516. bagging, val_score: 0.653850: 80%|######## | 8/10 [00:08<00:02, 1.11s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005318 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657371 valid's binary_logloss: 0.665888 [200] train's binary_logloss: 0.646844 valid's binary_logloss: 0.658962 [300] train's binary_logloss: 0.64144 valid's binary_logloss: 0.656351 [400] train's binary_logloss: 0.637614 valid's binary_logloss: 0.654985 [500] train's binary_logloss: 0.634424 valid's binary_logloss: 0.654861 [600] train's binary_logloss: 0.631665 valid's binary_logloss: 0.654356 [700] train's binary_logloss: 0.628979 valid's binary_logloss: 0.654652 Early stopping, best iteration is: [625] train's binary_logloss: 0.631017 valid's binary_logloss: 0.654177 bagging, val_score: 0.653850: 90%|######### | 9/10 [00:09<00:01, 1.19s/it][I 2020-09-27 04:41:50,003] Trial 35 finished with value: 0.6541765913434219 and parameters: {'bagging_fraction': 0.9914234683174175, 'bagging_freq': 7}. Best is trial 31 with value: 0.6538504332512516. bagging, val_score: 0.653850: 90%|######### | 9/10 [00:09<00:01, 1.19s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.006921 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657129 valid's binary_logloss: 0.666433 [200] train's binary_logloss: 0.646683 valid's binary_logloss: 0.658903 [300] train's binary_logloss: 0.641282 valid's binary_logloss: 0.656802 [400] train's binary_logloss: 0.637336 valid's binary_logloss: 0.655707 [500] train's binary_logloss: 0.634119 valid's binary_logloss: 0.654705 [600] train's binary_logloss: 0.631418 valid's binary_logloss: 0.654631 Early stopping, best iteration is: [510] train's binary_logloss: 0.633847 valid's binary_logloss: 0.654396 bagging, val_score: 0.653850: 100%|##########| 10/10 [00:11<00:00, 1.28s/it][I 2020-09-27 04:41:51,487] Trial 36 finished with value: 0.6543958028999971 and parameters: {'bagging_fraction': 0.9429983760903378, 'bagging_freq': 7}. Best is trial 31 with value: 0.6538504332512516. bagging, val_score: 0.653850: 100%|##########| 10/10 [00:11<00:00, 1.11s/it] feature_fraction_stage2, val_score: 0.653850: 0%| | 0/6 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000804 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657244 valid's binary_logloss: 0.665676 [200] train's binary_logloss: 0.646834 valid's binary_logloss: 0.657933 [300] train's binary_logloss: 0.641472 valid's binary_logloss: 0.655835 [400] train's binary_logloss: 0.637734 valid's binary_logloss: 0.654847 [500] train's binary_logloss: 0.634612 valid's binary_logloss: 0.654743 [600] train's binary_logloss: 0.631834 valid's binary_logloss: 0.654606 Early stopping, best iteration is: [559] train's binary_logloss: 0.632955 valid's binary_logloss: 0.654387 feature_fraction_stage2, val_score: 0.653850: 17%|#6 | 1/6 [00:01<00:05, 1.06s/it][I 2020-09-27 04:41:52,568] Trial 37 finished with value: 0.6543873568694546 and parameters: {'feature_fraction': 0.7520000000000001}. Best is trial 37 with value: 0.6543873568694546. feature_fraction_stage2, val_score: 0.653850: 17%|#6 | 1/6 [00:01<00:05, 1.06s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001058 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657307 valid's binary_logloss: 0.665041 [200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725 [300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506 [400] train's binary_logloss: 0.637619 valid's binary_logloss: 0.654377 [500] train's binary_logloss: 0.63434 valid's binary_logloss: 0.653884 Early stopping, best iteration is: [488] train's binary_logloss: 0.63474 valid's binary_logloss: 0.653747 feature_fraction_stage2, val_score: 0.653747: 33%|###3 | 2/6 [00:02<00:04, 1.04s/it][I 2020-09-27 04:41:53,565] Trial 38 finished with value: 0.6537470788868942 and parameters: {'feature_fraction': 0.7200000000000001}. Best is trial 38 with value: 0.6537470788868942. feature_fraction_stage2, val_score: 0.653747: 33%|###3 | 2/6 [00:02<00:04, 1.04s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001682 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657318 valid's binary_logloss: 0.665806 [200] train's binary_logloss: 0.646625 valid's binary_logloss: 0.658644 [300] train's binary_logloss: 0.641194 valid's binary_logloss: 0.656049 [400] train's binary_logloss: 0.63737 valid's binary_logloss: 0.655556 [500] train's binary_logloss: 0.634009 valid's binary_logloss: 0.654948 [600] train's binary_logloss: 0.63129 valid's binary_logloss: 0.65525 Early stopping, best iteration is: [557] train's binary_logloss: 0.63246 valid's binary_logloss: 0.654813 feature_fraction_stage2, val_score: 0.653747: 50%|##### | 3/6 [00:03<00:03, 1.21s/it][I 2020-09-27 04:41:55,150] Trial 39 finished with value: 0.6548128899258291 and parameters: {'feature_fraction': 0.8480000000000001}. Best is trial 38 with value: 0.6537470788868942. feature_fraction_stage2, val_score: 0.653747: 50%|##### | 3/6 [00:03<00:03, 1.21s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001085 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657207 valid's binary_logloss: 0.665659 [200] train's binary_logloss: 0.646755 valid's binary_logloss: 0.657975 [300] train's binary_logloss: 0.641365 valid's binary_logloss: 0.655752 [400] train's binary_logloss: 0.637538 valid's binary_logloss: 0.654478 [500] train's binary_logloss: 0.634276 valid's binary_logloss: 0.654404 [600] train's binary_logloss: 0.63152 valid's binary_logloss: 0.654509 Early stopping, best iteration is: [533] train's binary_logloss: 0.633346 valid's binary_logloss: 0.654136 feature_fraction_stage2, val_score: 0.653747: 67%|######6 | 4/6 [00:04<00:02, 1.22s/it][I 2020-09-27 04:41:56,409] Trial 40 finished with value: 0.6541357158243993 and parameters: {'feature_fraction': 0.88}. Best is trial 38 with value: 0.6537470788868942. feature_fraction_stage2, val_score: 0.653747: 67%|######6 | 4/6 [00:04<00:02, 1.22s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005676 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657297 valid's binary_logloss: 0.665521 [200] train's binary_logloss: 0.64689 valid's binary_logloss: 0.658315 [300] train's binary_logloss: 0.641484 valid's binary_logloss: 0.655464 [400] train's binary_logloss: 0.637638 valid's binary_logloss: 0.654412 [500] train's binary_logloss: 0.634281 valid's binary_logloss: 0.653965 [600] train's binary_logloss: 0.631536 valid's binary_logloss: 0.653942 Early stopping, best iteration is: [511] train's binary_logloss: 0.633982 valid's binary_logloss: 0.65385 feature_fraction_stage2, val_score: 0.653747: 83%|########3 | 5/6 [00:05<00:01, 1.18s/it][I 2020-09-27 04:41:57,501] Trial 41 finished with value: 0.6538504332512516 and parameters: {'feature_fraction': 0.8160000000000001}. Best is trial 38 with value: 0.6537470788868942. feature_fraction_stage2, val_score: 0.653747: 83%|########3 | 5/6 [00:06<00:01, 1.18s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010335 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657244 valid's binary_logloss: 0.665676 [200] train's binary_logloss: 0.646834 valid's binary_logloss: 0.657933 [300] train's binary_logloss: 0.641472 valid's binary_logloss: 0.655835 [400] train's binary_logloss: 0.637734 valid's binary_logloss: 0.654847 [500] train's binary_logloss: 0.634612 valid's binary_logloss: 0.654743 [600] train's binary_logloss: 0.631834 valid's binary_logloss: 0.654606 Early stopping, best iteration is: [559] train's binary_logloss: 0.632955 valid's binary_logloss: 0.654387 feature_fraction_stage2, val_score: 0.653747: 100%|##########| 6/6 [00:07<00:00, 1.29s/it][I 2020-09-27 04:41:59,040] Trial 42 finished with value: 0.6543873568694546 and parameters: {'feature_fraction': 0.784}. Best is trial 38 with value: 0.6537470788868942. feature_fraction_stage2, val_score: 0.653747: 100%|##########| 6/6 [00:07<00:00, 1.26s/it] regularization_factors, val_score: 0.653747: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000984 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657307 valid's binary_logloss: 0.665041 [200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725 [300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506 [400] train's binary_logloss: 0.637619 valid's binary_logloss: 0.654377 [500] train's binary_logloss: 0.63434 valid's binary_logloss: 0.653884 Early stopping, best iteration is: [488] train's binary_logloss: 0.63474 valid's binary_logloss: 0.653747 regularization_factors, val_score: 0.653747: 5%|5 | 1/20 [00:00<00:17, 1.11it/s][I 2020-09-27 04:41:59,955] Trial 43 finished with value: 0.6537470788623818 and parameters: {'lambda_l1': 1.164724543974479e-08, 'lambda_l2': 6.922005451695355e-06}. Best is trial 43 with value: 0.6537470788623818. regularization_factors, val_score: 0.653747: 5%|5 | 1/20 [00:00<00:17, 1.11it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000880 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657307 valid's binary_logloss: 0.665041 [200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725 [300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506 [400] train's binary_logloss: 0.637619 valid's binary_logloss: 0.654377 [500] train's binary_logloss: 0.63434 valid's binary_logloss: 0.653884 Early stopping, best iteration is: [488] train's binary_logloss: 0.63474 valid's binary_logloss: 0.653747 regularization_factors, val_score: 0.653747: 10%|# | 2/20 [00:01<00:16, 1.11it/s][I 2020-09-27 04:42:00,859] Trial 44 finished with value: 0.6537470788688746 and parameters: {'lambda_l1': 1.2101619670742207e-08, 'lambda_l2': 3.569458567905685e-06}. Best is trial 43 with value: 0.6537470788623818. regularization_factors, val_score: 0.653747: 10%|# | 2/20 [00:01<00:16, 1.11it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000804 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657307 valid's binary_logloss: 0.665041 [200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725 [300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506 [400] train's binary_logloss: 0.637619 valid's binary_logloss: 0.654377 [500] train's binary_logloss: 0.63434 valid's binary_logloss: 0.653884 Early stopping, best iteration is: [488] train's binary_logloss: 0.63474 valid's binary_logloss: 0.653747 regularization_factors, val_score: 0.653747: 15%|#5 | 3/20 [00:02<00:15, 1.08it/s][I 2020-09-27 04:42:01,825] Trial 45 finished with value: 0.653747078870228 and parameters: {'lambda_l1': 1.0924605897873326e-08, 'lambda_l2': 3.7438094337593657e-06}. Best is trial 43 with value: 0.6537470788623818. regularization_factors, val_score: 0.653747: 15%|#5 | 3/20 [00:02<00:15, 1.08it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009826 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657307 valid's binary_logloss: 0.665041 [200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725 [300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506 [400] train's binary_logloss: 0.637619 valid's binary_logloss: 0.654377 [500] train's binary_logloss: 0.63434 valid's binary_logloss: 0.653884 Early stopping, best iteration is: [488] train's binary_logloss: 0.63474 valid's binary_logloss: 0.653747 regularization_factors, val_score: 0.653747: 20%|## | 4/20 [00:04<00:20, 1.28s/it][I 2020-09-27 04:42:03,949] Trial 46 finished with value: 0.6537470788718757 and parameters: {'lambda_l1': 1.479258071574677e-08, 'lambda_l2': 3.0264187607450267e-06}. Best is trial 43 with value: 0.6537470788623818. regularization_factors, val_score: 0.653747: 20%|## | 4/20 [00:04<00:20, 1.28s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004927 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657307 valid's binary_logloss: 0.665041 [200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725 [300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506 [400] train's binary_logloss: 0.637619 valid's binary_logloss: 0.654377 [500] train's binary_logloss: 0.63434 valid's binary_logloss: 0.653884 Early stopping, best iteration is: [488] train's binary_logloss: 0.63474 valid's binary_logloss: 0.653747 regularization_factors, val_score: 0.653747: 25%|##5 | 5/20 [00:05<00:18, 1.20s/it][I 2020-09-27 04:42:04,966] Trial 47 finished with value: 0.6537470788672778 and parameters: {'lambda_l1': 1.2446439098514144e-08, 'lambda_l2': 4.099446197874459e-06}. Best is trial 43 with value: 0.6537470788623818. regularization_factors, val_score: 0.653747: 25%|##5 | 5/20 [00:05<00:18, 1.20s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004710 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657307 valid's binary_logloss: 0.665041 [200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725 [300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506 [400] train's binary_logloss: 0.637619 valid's binary_logloss: 0.654377 [500] train's binary_logloss: 0.63434 valid's binary_logloss: 0.653884 Early stopping, best iteration is: [488] train's binary_logloss: 0.63474 valid's binary_logloss: 0.653747 regularization_factors, val_score: 0.653747: 30%|### | 6/20 [00:06<00:15, 1.11s/it][I 2020-09-27 04:42:05,863] Trial 48 finished with value: 0.6537470788771779 and parameters: {'lambda_l1': 1.2455751430953037e-08, 'lambda_l2': 2.1195700656837565e-06}. Best is trial 43 with value: 0.6537470788623818. regularization_factors, val_score: 0.653747: 30%|### | 6/20 [00:06<00:15, 1.11s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000751 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657307 valid's binary_logloss: 0.665041 [200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725 [300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506 [400] train's binary_logloss: 0.637619 valid's binary_logloss: 0.654377 [500] train's binary_logloss: 0.63434 valid's binary_logloss: 0.653884 Early stopping, best iteration is: [488] train's binary_logloss: 0.63474 valid's binary_logloss: 0.653747 regularization_factors, val_score: 0.653747: 35%|###5 | 7/20 [00:08<00:15, 1.18s/it][I 2020-09-27 04:42:07,198] Trial 49 finished with value: 0.6537470788676445 and parameters: {'lambda_l1': 1.3371676440125825e-08, 'lambda_l2': 4.133315482813227e-06}. Best is trial 43 with value: 0.6537470788623818. regularization_factors, val_score: 0.653747: 35%|###5 | 7/20 [00:08<00:15, 1.18s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000868 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657307 valid's binary_logloss: 0.665041 [200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725 [300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506 [400] train's binary_logloss: 0.637619 valid's binary_logloss: 0.654377 [500] train's binary_logloss: 0.63434 valid's binary_logloss: 0.653884 Early stopping, best iteration is: [488] train's binary_logloss: 0.63474 valid's binary_logloss: 0.653747 regularization_factors, val_score: 0.653747: 40%|#### | 8/20 [00:09<00:13, 1.14s/it][I 2020-09-27 04:42:08,255] Trial 50 finished with value: 0.653747078859269 and parameters: {'lambda_l1': 1.1210362012517112e-08, 'lambda_l2': 6.439200774621681e-06}. Best is trial 50 with value: 0.653747078859269. regularization_factors, val_score: 0.653747: 40%|#### | 8/20 [00:09<00:13, 1.14s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000976 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657307 valid's binary_logloss: 0.665041 [200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725 [300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506 [400] train's binary_logloss: 0.637619 valid's binary_logloss: 0.654377 [500] train's binary_logloss: 0.63434 valid's binary_logloss: 0.653884 Early stopping, best iteration is: [488] train's binary_logloss: 0.63474 valid's binary_logloss: 0.653747 regularization_factors, val_score: 0.653747: 45%|####5 | 9/20 [00:10<00:11, 1.07s/it][I 2020-09-27 04:42:09,153] Trial 51 finished with value: 0.6537470788593895 and parameters: {'lambda_l1': 1.3454643930345611e-08, 'lambda_l2': 6.467158854009067e-06}. Best is trial 50 with value: 0.653747078859269. regularization_factors, val_score: 0.653747: 45%|####5 | 9/20 [00:10<00:11, 1.07s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.006696 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657307 valid's binary_logloss: 0.665041 [200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725 [300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506 [400] train's binary_logloss: 0.637619 valid's binary_logloss: 0.654377 [500] train's binary_logloss: 0.63434 valid's binary_logloss: 0.653884 Early stopping, best iteration is: [488] train's binary_logloss: 0.63474 valid's binary_logloss: 0.653747 regularization_factors, val_score: 0.653747: 50%|##### | 10/20 [00:11<00:10, 1.02s/it][I 2020-09-27 04:42:10,075] Trial 52 finished with value: 0.6537470788591189 and parameters: {'lambda_l1': 1.1646094395679663e-08, 'lambda_l2': 6.358551933697898e-06}. Best is trial 52 with value: 0.6537470788591189. regularization_factors, val_score: 0.653747: 50%|##### | 10/20 [00:11<00:10, 1.02s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011178 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657307 valid's binary_logloss: 0.665041 [200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725 [300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506 [400] train's binary_logloss: 0.637619 valid's binary_logloss: 0.654377 [500] train's binary_logloss: 0.63434 valid's binary_logloss: 0.653884 Early stopping, best iteration is: [488] train's binary_logloss: 0.63474 valid's binary_logloss: 0.653747 regularization_factors, val_score: 0.653747: 55%|#####5 | 11/20 [00:12<00:09, 1.11s/it][I 2020-09-27 04:42:11,387] Trial 53 finished with value: 0.6537470788474831 and parameters: {'lambda_l1': 1.0924146750147604e-08, 'lambda_l2': 1.2976298424529484e-05}. Best is trial 53 with value: 0.6537470788474831. regularization_factors, val_score: 0.653747: 55%|#####5 | 11/20 [00:12<00:09, 1.11s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005401 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657307 valid's binary_logloss: 0.665041 [200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725 [300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506 [400] train's binary_logloss: 0.637619 valid's binary_logloss: 0.654377 [500] train's binary_logloss: 0.63434 valid's binary_logloss: 0.653884 Early stopping, best iteration is: [488] train's binary_logloss: 0.63474 valid's binary_logloss: 0.653747 regularization_factors, val_score: 0.653747: 60%|###### | 12/20 [00:13<00:08, 1.05s/it][I 2020-09-27 04:42:12,300] Trial 54 finished with value: 0.65374707865709 and parameters: {'lambda_l1': 1.1771590372611887e-08, 'lambda_l2': 6.880400470984478e-05}. Best is trial 54 with value: 0.65374707865709. regularization_factors, val_score: 0.653747: 60%|###### | 12/20 [00:13<00:08, 1.05s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000845 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657307 valid's binary_logloss: 0.665041 [200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725 [300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506 [400] train's binary_logloss: 0.63762 valid's binary_logloss: 0.654377 [500] train's binary_logloss: 0.63434 valid's binary_logloss: 0.653884 Early stopping, best iteration is: [488] train's binary_logloss: 0.634741 valid's binary_logloss: 0.653747 regularization_factors, val_score: 0.653747: 65%|######5 | 13/20 [00:14<00:07, 1.02s/it][I 2020-09-27 04:42:13,235] Trial 55 finished with value: 0.6537470737830029 and parameters: {'lambda_l1': 1.2904541521956938e-08, 'lambda_l2': 0.0015325466166573717}. Best is trial 55 with value: 0.6537470737830029. regularization_factors, val_score: 0.653747: 65%|######5 | 13/20 [00:14<00:07, 1.02s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004533 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657308 valid's binary_logloss: 0.665041 [200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725 [300] train's binary_logloss: 0.64145 valid's binary_logloss: 0.65506 [400] train's binary_logloss: 0.637678 valid's binary_logloss: 0.654439 [500] train's binary_logloss: 0.634307 valid's binary_logloss: 0.653862 [600] train's binary_logloss: 0.631527 valid's binary_logloss: 0.653578 [700] train's binary_logloss: 0.628591 valid's binary_logloss: 0.653558 [800] train's binary_logloss: 0.625872 valid's binary_logloss: 0.653736 Early stopping, best iteration is: [750] train's binary_logloss: 0.627225 valid's binary_logloss: 0.653307 regularization_factors, val_score: 0.653307: 70%|####### | 14/20 [00:15<00:07, 1.22s/it][I 2020-09-27 04:42:14,917] Trial 56 finished with value: 0.653306938112237 and parameters: {'lambda_l1': 1.1130628894602006e-05, 'lambda_l2': 0.007158975373652711}. Best is trial 56 with value: 0.653306938112237. regularization_factors, val_score: 0.653307: 70%|####### | 14/20 [00:15<00:07, 1.22s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.019645 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657308 valid's binary_logloss: 0.665041 [200] train's binary_logloss: 0.646792 valid's binary_logloss: 0.657796 [300] train's binary_logloss: 0.641385 valid's binary_logloss: 0.655452 [400] train's binary_logloss: 0.637507 valid's binary_logloss: 0.654514 [500] train's binary_logloss: 0.634295 valid's binary_logloss: 0.653752 Early stopping, best iteration is: [490] train's binary_logloss: 0.63465 valid's binary_logloss: 0.653564 regularization_factors, val_score: 0.653307: 75%|#######5 | 15/20 [00:16<00:05, 1.15s/it][I 2020-09-27 04:42:15,922] Trial 57 finished with value: 0.6535639999340161 and parameters: {'lambda_l1': 0.00026873318136680526, 'lambda_l2': 0.011949459080105706}. Best is trial 56 with value: 0.653306938112237. regularization_factors, val_score: 0.653307: 75%|#######5 | 15/20 [00:16<00:05, 1.15s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000821 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657308 valid's binary_logloss: 0.665041 [200] train's binary_logloss: 0.646792 valid's binary_logloss: 0.657796 [300] train's binary_logloss: 0.641358 valid's binary_logloss: 0.655395 [400] train's binary_logloss: 0.637525 valid's binary_logloss: 0.654849 [500] train's binary_logloss: 0.63422 valid's binary_logloss: 0.654174 [600] train's binary_logloss: 0.631488 valid's binary_logloss: 0.654245 Early stopping, best iteration is: [588] train's binary_logloss: 0.631831 valid's binary_logloss: 0.654014 regularization_factors, val_score: 0.653307: 80%|######## | 16/20 [00:17<00:04, 1.12s/it][I 2020-09-27 04:42:16,965] Trial 58 finished with value: 0.6540142566432265 and parameters: {'lambda_l1': 0.0002391919617921031, 'lambda_l2': 0.020184808500303845}. Best is trial 56 with value: 0.653306938112237. regularization_factors, val_score: 0.653307: 80%|######## | 16/20 [00:17<00:04, 1.12s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004897 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657308 valid's binary_logloss: 0.665041 [200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725 [300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506 [400] train's binary_logloss: 0.63762 valid's binary_logloss: 0.654377 [500] train's binary_logloss: 0.634341 valid's binary_logloss: 0.653884 Early stopping, best iteration is: [488] train's binary_logloss: 0.634741 valid's binary_logloss: 0.653747 regularization_factors, val_score: 0.653307: 85%|########5 | 17/20 [00:18<00:03, 1.06s/it][I 2020-09-27 04:42:17,888] Trial 59 finished with value: 0.6537470700910835 and parameters: {'lambda_l1': 0.00026680911329737025, 'lambda_l2': 0.0033110276633739268}. Best is trial 56 with value: 0.653306938112237. regularization_factors, val_score: 0.653307: 85%|########5 | 17/20 [00:18<00:03, 1.06s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004895 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657308 valid's binary_logloss: 0.665041 [200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725 [300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506 [400] train's binary_logloss: 0.637621 valid's binary_logloss: 0.654377 [500] train's binary_logloss: 0.634341 valid's binary_logloss: 0.653884 Early stopping, best iteration is: [488] train's binary_logloss: 0.634742 valid's binary_logloss: 0.653747 regularization_factors, val_score: 0.653307: 90%|######### | 18/20 [00:20<00:02, 1.17s/it][I 2020-09-27 04:42:19,323] Trial 60 finished with value: 0.6537470677716328 and parameters: {'lambda_l1': 0.0003014398725480866, 'lambda_l2': 0.004090143660649428}. Best is trial 56 with value: 0.653306938112237. regularization_factors, val_score: 0.653307: 90%|######### | 18/20 [00:20<00:02, 1.17s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000797 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657308 valid's binary_logloss: 0.665041 [200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725 [300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506 [400] train's binary_logloss: 0.637621 valid's binary_logloss: 0.654377 [500] train's binary_logloss: 0.634342 valid's binary_logloss: 0.653884 Early stopping, best iteration is: [488] train's binary_logloss: 0.634742 valid's binary_logloss: 0.653747 regularization_factors, val_score: 0.653307: 95%|#########5| 19/20 [00:21<00:01, 1.10s/it][I 2020-09-27 04:42:20,258] Trial 61 finished with value: 0.6537470687771331 and parameters: {'lambda_l1': 0.0005632008256227812, 'lambda_l2': 0.004446124070486416}. Best is trial 56 with value: 0.653306938112237. regularization_factors, val_score: 0.653307: 95%|#########5| 19/20 [00:21<00:01, 1.10s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000827 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657308 valid's binary_logloss: 0.665041 [200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725 [300] train's binary_logloss: 0.641449 valid's binary_logloss: 0.65506 [400] train's binary_logloss: 0.637621 valid's binary_logloss: 0.654377 [500] train's binary_logloss: 0.634341 valid's binary_logloss: 0.653884 Early stopping, best iteration is: [488] train's binary_logloss: 0.634742 valid's binary_logloss: 0.653747 regularization_factors, val_score: 0.653307: 100%|##########| 20/20 [00:22<00:00, 1.04s/it][I 2020-09-27 04:42:21,158] Trial 62 finished with value: 0.6537470665554331 and parameters: {'lambda_l1': 0.0003132915633232416, 'lambda_l2': 0.004489563552243147}. Best is trial 56 with value: 0.653306938112237. regularization_factors, val_score: 0.653307: 100%|##########| 20/20 [00:22<00:00, 1.11s/it] min_data_in_leaf, val_score: 0.653307: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004952 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657308 valid's binary_logloss: 0.665041 [200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725 [300] train's binary_logloss: 0.641484 valid's binary_logloss: 0.655571 [400] train's binary_logloss: 0.63758 valid's binary_logloss: 0.654925 Early stopping, best iteration is: [368] train's binary_logloss: 0.638713 valid's binary_logloss: 0.654841 min_data_in_leaf, val_score: 0.653307: 20%|## | 1/5 [00:00<00:03, 1.21it/s][I 2020-09-27 04:42:21,996] Trial 63 finished with value: 0.6548408998104529 and parameters: {'min_child_samples': 5}. Best is trial 63 with value: 0.6548408998104529. min_data_in_leaf, val_score: 0.653307: 20%|## | 1/5 [00:00<00:03, 1.21it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005026 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657308 valid's binary_logloss: 0.665041 [200] train's binary_logloss: 0.646871 valid's binary_logloss: 0.657816 [300] train's binary_logloss: 0.641553 valid's binary_logloss: 0.655747 [400] train's binary_logloss: 0.637618 valid's binary_logloss: 0.655166 [500] train's binary_logloss: 0.63433 valid's binary_logloss: 0.654821 Early stopping, best iteration is: [483] train's binary_logloss: 0.634862 valid's binary_logloss: 0.654649 min_data_in_leaf, val_score: 0.653307: 40%|#### | 2/5 [00:02<00:02, 1.04it/s][I 2020-09-27 04:42:23,270] Trial 64 finished with value: 0.6546487352338517 and parameters: {'min_child_samples': 25}. Best is trial 64 with value: 0.6546487352338517. min_data_in_leaf, val_score: 0.653307: 40%|#### | 2/5 [00:02<00:02, 1.04it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004979 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657308 valid's binary_logloss: 0.665041 [200] train's binary_logloss: 0.646783 valid's binary_logloss: 0.657725 [300] train's binary_logloss: 0.641452 valid's binary_logloss: 0.655218 [400] train's binary_logloss: 0.637601 valid's binary_logloss: 0.654481 [500] train's binary_logloss: 0.634142 valid's binary_logloss: 0.653967 [600] train's binary_logloss: 0.631213 valid's binary_logloss: 0.654248 Early stopping, best iteration is: [539] train's binary_logloss: 0.632957 valid's binary_logloss: 0.653768 min_data_in_leaf, val_score: 0.653307: 60%|###### | 3/5 [00:03<00:01, 1.02it/s][I 2020-09-27 04:42:24,298] Trial 65 finished with value: 0.653767937805501 and parameters: {'min_child_samples': 10}. Best is trial 65 with value: 0.653767937805501. min_data_in_leaf, val_score: 0.653307: 60%|###### | 3/5 [00:03<00:01, 1.02it/s][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000881 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657308 valid's binary_logloss: 0.665041 [200] train's binary_logloss: 0.646871 valid's binary_logloss: 0.657816 [300] train's binary_logloss: 0.641452 valid's binary_logloss: 0.655693 [400] train's binary_logloss: 0.637718 valid's binary_logloss: 0.654576 [500] train's binary_logloss: 0.634473 valid's binary_logloss: 0.654212 [600] train's binary_logloss: 0.631693 valid's binary_logloss: 0.654007 Early stopping, best iteration is: [580] train's binary_logloss: 0.63221 valid's binary_logloss: 0.653668 min_data_in_leaf, val_score: 0.653307: 80%|######## | 4/5 [00:04<00:01, 1.01s/it][I 2020-09-27 04:42:25,378] Trial 66 finished with value: 0.6536677236980378 and parameters: {'min_child_samples': 50}. Best is trial 66 with value: 0.6536677236980378. min_data_in_leaf, val_score: 0.653307: 80%|######## | 4/5 [00:04<00:01, 1.01s/it][LightGBM] [Info] Number of positive: 13145, number of negative: 12854 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004998 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505596 -> initscore=0.022386 [LightGBM] [Info] Start training from score 0.022386 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657308 valid's binary_logloss: 0.665041 [200] train's binary_logloss: 0.646872 valid's binary_logloss: 0.657799 [300] train's binary_logloss: 0.641589 valid's binary_logloss: 0.655365 [400] train's binary_logloss: 0.638214 valid's binary_logloss: 0.65449 [500] train's binary_logloss: 0.635191 valid's binary_logloss: 0.653777 [600] train's binary_logloss: 0.632659 valid's binary_logloss: 0.653578 Early stopping, best iteration is: [576] train's binary_logloss: 0.633154 valid's binary_logloss: 0.653401 min_data_in_leaf, val_score: 0.653307: 100%|##########| 5/5 [00:05<00:00, 1.17s/it][I 2020-09-27 04:42:26,912] Trial 67 finished with value: 0.653400524735626 and parameters: {'min_child_samples': 100}. Best is trial 67 with value: 0.653400524735626. min_data_in_leaf, val_score: 0.653307: 100%|##########| 5/5 [00:05<00:00, 1.15s/it] Fold : 4 [I 2020-09-27 04:42:26,996] A new study created in memory with name: no-name-95b73563-c9f0-498a-9add-fe0ff51242f2 feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000967 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.569629 valid's binary_logloss: 0.666525 Early stopping, best iteration is: [63] train's binary_logloss: 0.595461 valid's binary_logloss: 0.665516 feature_fraction, val_score: 0.665516: 14%|#4 | 1/7 [00:00<00:03, 1.75it/s][I 2020-09-27 04:42:27,577] Trial 0 finished with value: 0.6655156235172205 and parameters: {'feature_fraction': 1.0}. Best is trial 0 with value: 0.6655156235172205. feature_fraction, val_score: 0.665516: 14%|#4 | 1/7 [00:00<00:03, 1.75it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000965 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.572137 valid's binary_logloss: 0.663019 Early stopping, best iteration is: [48] train's binary_logloss: 0.610212 valid's binary_logloss: 0.661166 feature_fraction, val_score: 0.661166: 29%|##8 | 2/7 [00:01<00:02, 1.85it/s][I 2020-09-27 04:42:28,047] Trial 1 finished with value: 0.6611663859231157 and parameters: {'feature_fraction': 0.8}. Best is trial 1 with value: 0.6611663859231157. feature_fraction, val_score: 0.661166: 29%|##8 | 2/7 [00:01<00:02, 1.85it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004690 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.57535 valid's binary_logloss: 0.662198 Early stopping, best iteration is: [58] train's binary_logloss: 0.604866 valid's binary_logloss: 0.660561 feature_fraction, val_score: 0.660561: 43%|####2 | 3/7 [00:01<00:02, 1.92it/s][I 2020-09-27 04:42:28,523] Trial 2 finished with value: 0.660560693860469 and parameters: {'feature_fraction': 0.6}. Best is trial 2 with value: 0.660560693860469. feature_fraction, val_score: 0.660561: 43%|####2 | 3/7 [00:01<00:02, 1.92it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000938 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.571052 valid's binary_logloss: 0.6616 Early stopping, best iteration is: [50] train's binary_logloss: 0.607427 valid's binary_logloss: 0.660867 feature_fraction, val_score: 0.660561: 57%|#####7 | 4/7 [00:02<00:01, 1.94it/s][I 2020-09-27 04:42:29,026] Trial 3 finished with value: 0.6608672257275293 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 2 with value: 0.660560693860469. feature_fraction, val_score: 0.660561: 57%|#####7 | 4/7 [00:02<00:01, 1.94it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000868 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.574313 valid's binary_logloss: 0.662136 Early stopping, best iteration is: [97] train's binary_logloss: 0.576201 valid's binary_logloss: 0.661927 feature_fraction, val_score: 0.660561: 71%|#######1 | 5/7 [00:02<00:01, 1.79it/s][I 2020-09-27 04:42:29,690] Trial 4 finished with value: 0.661927284753647 and parameters: {'feature_fraction': 0.7}. Best is trial 2 with value: 0.660560693860469. feature_fraction, val_score: 0.660561: 71%|#######1 | 5/7 [00:02<00:01, 1.79it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012827 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.578468 valid's binary_logloss: 0.661585 Early stopping, best iteration is: [54] train's binary_logloss: 0.610009 valid's binary_logloss: 0.661106 feature_fraction, val_score: 0.660561: 86%|########5 | 6/7 [00:03<00:00, 1.58it/s][I 2020-09-27 04:42:30,492] Trial 5 finished with value: 0.6611057071152201 and parameters: {'feature_fraction': 0.5}. Best is trial 2 with value: 0.660560693860469. feature_fraction, val_score: 0.660561: 86%|########5 | 6/7 [00:03<00:00, 1.58it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000508 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.581603 valid's binary_logloss: 0.660601 Early stopping, best iteration is: [81] train's binary_logloss: 0.593654 valid's binary_logloss: 0.660443 feature_fraction, val_score: 0.660443: 100%|##########| 7/7 [00:03<00:00, 1.68it/s][I 2020-09-27 04:42:30,996] Trial 6 finished with value: 0.6604425614247917 and parameters: {'feature_fraction': 0.4}. Best is trial 6 with value: 0.6604425614247917. feature_fraction, val_score: 0.660443: 100%|##########| 7/7 [00:04<00:00, 1.75it/s] num_leaves, val_score: 0.660443: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004658 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.356495 valid's binary_logloss: 0.671539 Early stopping, best iteration is: [41] train's binary_logloss: 0.499664 valid's binary_logloss: 0.664498 num_leaves, val_score: 0.660443: 5%|5 | 1/20 [00:00<00:16, 1.14it/s][I 2020-09-27 04:42:31,893] Trial 7 finished with value: 0.6644984899613434 and parameters: {'num_leaves': 190}. Best is trial 7 with value: 0.6644984899613434. num_leaves, val_score: 0.660443: 5%|5 | 1/20 [00:00<00:16, 1.14it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000609 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.526375 valid's binary_logloss: 0.665053 Early stopping, best iteration is: [67] train's binary_logloss: 0.561219 valid's binary_logloss: 0.662832 num_leaves, val_score: 0.660443: 10%|# | 2/20 [00:01<00:14, 1.28it/s][I 2020-09-27 04:42:32,449] Trial 8 finished with value: 0.6628317798276503 and parameters: {'num_leaves': 60}. Best is trial 8 with value: 0.6628317798276503. num_leaves, val_score: 0.660443: 10%|# | 2/20 [00:01<00:14, 1.28it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000379 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.608789 valid's binary_logloss: 0.659931 [200] train's binary_logloss: 0.573948 valid's binary_logloss: 0.661564 Early stopping, best iteration is: [106] train's binary_logloss: 0.606284 valid's binary_logloss: 0.659493 num_leaves, val_score: 0.659493: 15%|#5 | 3/20 [00:01<00:11, 1.48it/s][I 2020-09-27 04:42:32,881] Trial 9 finished with value: 0.6594930122196023 and parameters: {'num_leaves': 19}. Best is trial 9 with value: 0.6594930122196023. num_leaves, val_score: 0.659493: 15%|#5 | 3/20 [00:01<00:11, 1.48it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000374 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.62851 valid's binary_logloss: 0.661694 [200] train's binary_logloss: 0.606786 valid's binary_logloss: 0.660582 Early stopping, best iteration is: [177] train's binary_logloss: 0.611571 valid's binary_logloss: 0.660135 num_leaves, val_score: 0.659493: 20%|## | 4/20 [00:02<00:10, 1.56it/s][I 2020-09-27 04:42:33,444] Trial 10 finished with value: 0.6601347890202464 and parameters: {'num_leaves': 11}. Best is trial 9 with value: 0.6594930122196023. num_leaves, val_score: 0.659493: 20%|## | 4/20 [00:02<00:10, 1.56it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002220 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.623153 valid's binary_logloss: 0.659612 [200] train's binary_logloss: 0.598247 valid's binary_logloss: 0.660773 Early stopping, best iteration is: [106] train's binary_logloss: 0.621362 valid's binary_logloss: 0.659491 num_leaves, val_score: 0.659491: 25%|##5 | 5/20 [00:03<00:10, 1.41it/s][I 2020-09-27 04:42:34,315] Trial 11 finished with value: 0.6594905167648981 and parameters: {'num_leaves': 13}. Best is trial 11 with value: 0.6594905167648981. num_leaves, val_score: 0.659491: 25%|##5 | 5/20 [00:03<00:10, 1.41it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000240 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.633624 valid's binary_logloss: 0.661805 [200] train's binary_logloss: 0.6154 valid's binary_logloss: 0.661626 Early stopping, best iteration is: [118] train's binary_logloss: 0.629667 valid's binary_logloss: 0.661053 num_leaves, val_score: 0.659491: 30%|### | 6/20 [00:03<00:08, 1.61it/s][I 2020-09-27 04:42:34,733] Trial 12 finished with value: 0.6610531437027433 and parameters: {'num_leaves': 9}. Best is trial 11 with value: 0.6594905167648981. num_leaves, val_score: 0.659491: 30%|### | 6/20 [00:03<00:08, 1.61it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000443 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.489932 valid's binary_logloss: 0.664356 Early stopping, best iteration is: [42] train's binary_logloss: 0.574496 valid's binary_logloss: 0.661322 num_leaves, val_score: 0.659491: 35%|###5 | 7/20 [00:04<00:07, 1.63it/s][I 2020-09-27 04:42:35,325] Trial 13 finished with value: 0.6613221980695407 and parameters: {'num_leaves': 83}. Best is trial 11 with value: 0.6594905167648981. num_leaves, val_score: 0.659491: 35%|###5 | 7/20 [00:04<00:07, 1.63it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000456 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.668466 valid's binary_logloss: 0.674773 [200] train's binary_logloss: 0.658603 valid's binary_logloss: 0.667149 [300] train's binary_logloss: 0.653224 valid's binary_logloss: 0.663743 [400] train's binary_logloss: 0.649988 valid's binary_logloss: 0.661832 [500] train's binary_logloss: 0.647929 valid's binary_logloss: 0.661043 [600] train's binary_logloss: 0.646563 valid's binary_logloss: 0.660633 [700] train's binary_logloss: 0.645608 valid's binary_logloss: 0.660633 [800] train's binary_logloss: 0.644889 valid's binary_logloss: 0.660438 [900] train's binary_logloss: 0.644316 valid's binary_logloss: 0.66034 [1000] train's binary_logloss: 0.64384 valid's binary_logloss: 0.660316 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.64384 valid's binary_logloss: 0.660316 num_leaves, val_score: 0.659491: 40%|#### | 8/20 [00:05<00:09, 1.28it/s][I 2020-09-27 04:42:36,492] Trial 14 finished with value: 0.6603164467173235 and parameters: {'num_leaves': 2}. Best is trial 11 with value: 0.6594905167648981. num_leaves, val_score: 0.659491: 40%|#### | 8/20 [00:05<00:09, 1.28it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000395 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.481541 valid's binary_logloss: 0.666981 Early stopping, best iteration is: [44] train's binary_logloss: 0.566666 valid's binary_logloss: 0.662735 num_leaves, val_score: 0.659491: 45%|####5 | 9/20 [00:06<00:08, 1.35it/s][I 2020-09-27 04:42:37,143] Trial 15 finished with value: 0.6627351233501201 and parameters: {'num_leaves': 88}. Best is trial 11 with value: 0.6594905167648981. num_leaves, val_score: 0.659491: 45%|####5 | 9/20 [00:06<00:08, 1.35it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000349 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.3973 valid's binary_logloss: 0.672802 Early stopping, best iteration is: [29] train's binary_logloss: 0.560753 valid's binary_logloss: 0.66553 num_leaves, val_score: 0.659491: 50%|##### | 10/20 [00:07<00:09, 1.08it/s][I 2020-09-27 04:42:38,498] Trial 16 finished with value: 0.6655299653769075 and parameters: {'num_leaves': 152}. Best is trial 11 with value: 0.6594905167648981. num_leaves, val_score: 0.659491: 50%|##### | 10/20 [00:07<00:09, 1.08it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000373 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.299677 valid's binary_logloss: 0.679053 Early stopping, best iteration is: [28] train's binary_logloss: 0.514613 valid's binary_logloss: 0.667796 num_leaves, val_score: 0.659491: 55%|#####5 | 11/20 [00:08<00:08, 1.03it/s][I 2020-09-27 04:42:39,559] Trial 17 finished with value: 0.6677957520143049 and parameters: {'num_leaves': 256}. Best is trial 11 with value: 0.6594905167648981. num_leaves, val_score: 0.659491: 55%|#####5 | 11/20 [00:08<00:08, 1.03it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000315 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.5506 valid's binary_logloss: 0.661771 Early stopping, best iteration is: [46] train's binary_logloss: 0.602349 valid's binary_logloss: 0.660582 num_leaves, val_score: 0.659491: 60%|###### | 12/20 [00:08<00:06, 1.23it/s][I 2020-09-27 04:42:40,010] Trial 18 finished with value: 0.6605823886498043 and parameters: {'num_leaves': 47}. Best is trial 11 with value: 0.6594905167648981. num_leaves, val_score: 0.659491: 60%|###### | 12/20 [00:09<00:06, 1.23it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000360 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.573596 valid's binary_logloss: 0.659418 Early stopping, best iteration is: [95] train's binary_logloss: 0.576854 valid's binary_logloss: 0.658933 num_leaves, val_score: 0.658933: 65%|######5 | 13/20 [00:09<00:04, 1.40it/s][I 2020-09-27 04:42:40,490] Trial 19 finished with value: 0.6589331658992789 and parameters: {'num_leaves': 35}. Best is trial 19 with value: 0.6589331658992789. num_leaves, val_score: 0.658933: 65%|######5 | 13/20 [00:09<00:04, 1.40it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000370 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.433641 valid's binary_logloss: 0.669498 Early stopping, best iteration is: [30] train's binary_logloss: 0.5737 valid's binary_logloss: 0.666057 num_leaves, val_score: 0.658933: 70%|####### | 14/20 [00:10<00:04, 1.31it/s][I 2020-09-27 04:42:41,370] Trial 20 finished with value: 0.666056567776029 and parameters: {'num_leaves': 123}. Best is trial 19 with value: 0.6589331658992789. num_leaves, val_score: 0.658933: 70%|####### | 14/20 [00:10<00:04, 1.31it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011203 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.578104 valid's binary_logloss: 0.662017 Early stopping, best iteration is: [72] train's binary_logloss: 0.596911 valid's binary_logloss: 0.661574 num_leaves, val_score: 0.658933: 75%|#######5 | 15/20 [00:10<00:03, 1.41it/s][I 2020-09-27 04:42:41,958] Trial 21 finished with value: 0.6615737018014829 and parameters: {'num_leaves': 33}. Best is trial 19 with value: 0.6589331658992789. num_leaves, val_score: 0.658933: 75%|#######5 | 15/20 [00:10<00:03, 1.41it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000381 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.668466 valid's binary_logloss: 0.674773 [200] train's binary_logloss: 0.658603 valid's binary_logloss: 0.667149 [300] train's binary_logloss: 0.653224 valid's binary_logloss: 0.663743 [400] train's binary_logloss: 0.649988 valid's binary_logloss: 0.661832 [500] train's binary_logloss: 0.647929 valid's binary_logloss: 0.661043 [600] train's binary_logloss: 0.646563 valid's binary_logloss: 0.660633 [700] train's binary_logloss: 0.645608 valid's binary_logloss: 0.660633 [800] train's binary_logloss: 0.644889 valid's binary_logloss: 0.660438 [900] train's binary_logloss: 0.644316 valid's binary_logloss: 0.66034 [1000] train's binary_logloss: 0.64384 valid's binary_logloss: 0.660316 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.64384 valid's binary_logloss: 0.660316 num_leaves, val_score: 0.658933: 80%|######## | 16/20 [00:12<00:03, 1.18it/s][I 2020-09-27 04:42:43,132] Trial 22 finished with value: 0.6603164467173235 and parameters: {'num_leaves': 2}. Best is trial 19 with value: 0.6589331658992789. num_leaves, val_score: 0.658933: 80%|######## | 16/20 [00:12<00:03, 1.18it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000388 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.581603 valid's binary_logloss: 0.660601 Early stopping, best iteration is: [81] train's binary_logloss: 0.593654 valid's binary_logloss: 0.660443 num_leaves, val_score: 0.658933: 85%|########5 | 17/20 [00:12<00:02, 1.34it/s][I 2020-09-27 04:42:43,636] Trial 23 finished with value: 0.6604425614247916 and parameters: {'num_leaves': 31}. Best is trial 19 with value: 0.6589331658992789. num_leaves, val_score: 0.658933: 85%|########5 | 17/20 [00:12<00:02, 1.34it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000430 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.517038 valid's binary_logloss: 0.664708 Early stopping, best iteration is: [42] train's binary_logloss: 0.589858 valid's binary_logloss: 0.660965 num_leaves, val_score: 0.658933: 90%|######### | 18/20 [00:13<00:01, 1.47it/s][I 2020-09-27 04:42:44,171] Trial 24 finished with value: 0.660965322744683 and parameters: {'num_leaves': 66}. Best is trial 19 with value: 0.6589331658992789. num_leaves, val_score: 0.658933: 90%|######### | 18/20 [00:13<00:01, 1.47it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000404 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.590111 valid's binary_logloss: 0.659776 Early stopping, best iteration is: [96] train's binary_logloss: 0.592038 valid's binary_logloss: 0.65953 num_leaves, val_score: 0.658933: 95%|#########5| 19/20 [00:13<00:00, 1.61it/s][I 2020-09-27 04:42:44,655] Trial 25 finished with value: 0.6595301896319116 and parameters: {'num_leaves': 27}. Best is trial 19 with value: 0.6589331658992789. num_leaves, val_score: 0.658933: 95%|#########5| 19/20 [00:13<00:00, 1.61it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000349 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.454148 valid's binary_logloss: 0.667807 Early stopping, best iteration is: [45] train's binary_logloss: 0.54925 valid's binary_logloss: 0.664334 num_leaves, val_score: 0.658933: 100%|##########| 20/20 [00:14<00:00, 1.28it/s][I 2020-09-27 04:42:45,813] Trial 26 finished with value: 0.6643342584470427 and parameters: {'num_leaves': 106}. Best is trial 19 with value: 0.6589331658992789. num_leaves, val_score: 0.658933: 100%|##########| 20/20 [00:14<00:00, 1.35it/s] bagging, val_score: 0.658933: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000378 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.57621 valid's binary_logloss: 0.663844 Early stopping, best iteration is: [82] train's binary_logloss: 0.587864 valid's binary_logloss: 0.662968 bagging, val_score: 0.658933: 10%|# | 1/10 [00:00<00:04, 1.88it/s][I 2020-09-27 04:42:46,361] Trial 27 finished with value: 0.6629683704520333 and parameters: {'bagging_fraction': 0.6524827037560431, 'bagging_freq': 6}. Best is trial 27 with value: 0.6629683704520333. bagging, val_score: 0.658933: 10%|# | 1/10 [00:00<00:04, 1.88it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000416 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.573429 valid's binary_logloss: 0.66344 Early stopping, best iteration is: [70] train's binary_logloss: 0.595186 valid's binary_logloss: 0.661013 bagging, val_score: 0.658933: 20%|## | 2/10 [00:00<00:04, 1.96it/s][I 2020-09-27 04:42:46,824] Trial 28 finished with value: 0.6610133814310396 and parameters: {'bagging_fraction': 0.9909619380758035, 'bagging_freq': 1}. Best is trial 28 with value: 0.6610133814310396. bagging, val_score: 0.658933: 20%|## | 2/10 [00:01<00:04, 1.96it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000474 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.580089 valid's binary_logloss: 0.668617 Early stopping, best iteration is: [39] train's binary_logloss: 0.625728 valid's binary_logloss: 0.665767 bagging, val_score: 0.658933: 30%|### | 3/10 [00:01<00:03, 2.09it/s][I 2020-09-27 04:42:47,229] Trial 29 finished with value: 0.6657670269268177 and parameters: {'bagging_fraction': 0.42763463150211456, 'bagging_freq': 1}. Best is trial 28 with value: 0.6610133814310396. bagging, val_score: 0.658933: 30%|### | 3/10 [00:01<00:03, 2.09it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000410 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.573775 valid's binary_logloss: 0.662206 Early stopping, best iteration is: [84] train's binary_logloss: 0.585244 valid's binary_logloss: 0.661816 bagging, val_score: 0.658933: 40%|#### | 4/10 [00:01<00:02, 2.03it/s][I 2020-09-27 04:42:47,755] Trial 30 finished with value: 0.6618156054088046 and parameters: {'bagging_fraction': 0.9647411027897097, 'bagging_freq': 7}. Best is trial 28 with value: 0.6610133814310396. bagging, val_score: 0.658933: 40%|#### | 4/10 [00:01<00:02, 2.03it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004478 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.579254 valid's binary_logloss: 0.665041 Early stopping, best iteration is: [60] train's binary_logloss: 0.607077 valid's binary_logloss: 0.658928 bagging, val_score: 0.658928: 50%|##### | 5/10 [00:02<00:02, 2.12it/s][I 2020-09-27 04:42:48,176] Trial 31 finished with value: 0.6589276849000186 and parameters: {'bagging_fraction': 0.4607699724160956, 'bagging_freq': 4}. Best is trial 31 with value: 0.6589276849000186. bagging, val_score: 0.658928: 50%|##### | 5/10 [00:02<00:02, 2.12it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000381 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.580916 valid's binary_logloss: 0.671067 Early stopping, best iteration is: [36] train's binary_logloss: 0.628471 valid's binary_logloss: 0.662306 bagging, val_score: 0.658928: 60%|###### | 6/10 [00:03<00:02, 1.82it/s][I 2020-09-27 04:42:48,914] Trial 32 finished with value: 0.6623061121182913 and parameters: {'bagging_fraction': 0.4089026397973168, 'bagging_freq': 4}. Best is trial 31 with value: 0.6589276849000186. bagging, val_score: 0.658928: 60%|###### | 6/10 [00:03<00:02, 1.82it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001551 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.57696 valid's binary_logloss: 0.664792 Early stopping, best iteration is: [44] train's binary_logloss: 0.620171 valid's binary_logloss: 0.661311 bagging, val_score: 0.658928: 70%|####### | 7/10 [00:03<00:01, 1.75it/s][I 2020-09-27 04:42:49,530] Trial 33 finished with value: 0.6613113079015099 and parameters: {'bagging_fraction': 0.5861812306541639, 'bagging_freq': 4}. Best is trial 31 with value: 0.6589276849000186. bagging, val_score: 0.658928: 70%|####### | 7/10 [00:03<00:01, 1.75it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000437 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.575286 valid's binary_logloss: 0.663569 Early stopping, best iteration is: [62] train's binary_logloss: 0.60231 valid's binary_logloss: 0.661403 bagging, val_score: 0.658928: 80%|######## | 8/10 [00:04<00:01, 1.80it/s][I 2020-09-27 04:42:50,048] Trial 34 finished with value: 0.6614028699158252 and parameters: {'bagging_fraction': 0.8070642452561286, 'bagging_freq': 3}. Best is trial 31 with value: 0.6589276849000186. bagging, val_score: 0.658928: 80%|######## | 8/10 [00:04<00:01, 1.80it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000238 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.578059 valid's binary_logloss: 0.66561 Early stopping, best iteration is: [57] train's binary_logloss: 0.60883 valid's binary_logloss: 0.660744 bagging, val_score: 0.658928: 90%|######### | 9/10 [00:04<00:00, 1.85it/s][I 2020-09-27 04:42:50,558] Trial 35 finished with value: 0.6607440151387087 and parameters: {'bagging_fraction': 0.5229694779889069, 'bagging_freq': 5}. Best is trial 31 with value: 0.6589276849000186. bagging, val_score: 0.658928: 90%|######### | 9/10 [00:04<00:00, 1.85it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000982 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.574809 valid's binary_logloss: 0.665725 Early stopping, best iteration is: [73] train's binary_logloss: 0.593617 valid's binary_logloss: 0.664242 bagging, val_score: 0.658928: 100%|##########| 10/10 [00:05<00:00, 1.76it/s][I 2020-09-27 04:42:51,184] Trial 36 finished with value: 0.6642418801252681 and parameters: {'bagging_fraction': 0.7857998781191295, 'bagging_freq': 2}. Best is trial 31 with value: 0.6589276849000186. bagging, val_score: 0.658928: 100%|##########| 10/10 [00:05<00:00, 1.86it/s] feature_fraction_stage2, val_score: 0.658928: 0%| | 0/3 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004751 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.576158 valid's binary_logloss: 0.667365 Early stopping, best iteration is: [40] train's binary_logloss: 0.622262 valid's binary_logloss: 0.662845 feature_fraction_stage2, val_score: 0.658928: 33%|###3 | 1/3 [00:00<00:00, 2.32it/s][I 2020-09-27 04:42:51,633] Trial 37 finished with value: 0.6628451796786387 and parameters: {'feature_fraction': 0.44800000000000006}. Best is trial 37 with value: 0.6628451796786387. feature_fraction_stage2, val_score: 0.658928: 33%|###3 | 1/3 [00:00<00:00, 2.32it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000343 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.577968 valid's binary_logloss: 0.668202 Early stopping, best iteration is: [34] train's binary_logloss: 0.629452 valid's binary_logloss: 0.662397 feature_fraction_stage2, val_score: 0.658928: 67%|######6 | 2/3 [00:00<00:00, 2.34it/s][I 2020-09-27 04:42:52,051] Trial 38 finished with value: 0.6623967977826815 and parameters: {'feature_fraction': 0.41600000000000004}. Best is trial 38 with value: 0.6623967977826815. feature_fraction_stage2, val_score: 0.658928: 67%|######6 | 2/3 [00:00<00:00, 2.34it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007814 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.576158 valid's binary_logloss: 0.667365 Early stopping, best iteration is: [40] train's binary_logloss: 0.622262 valid's binary_logloss: 0.662845 feature_fraction_stage2, val_score: 0.658928: 100%|##########| 3/3 [00:01<00:00, 2.00it/s][I 2020-09-27 04:42:52,719] Trial 39 finished with value: 0.6628451796786387 and parameters: {'feature_fraction': 0.48000000000000004}. Best is trial 38 with value: 0.6623967977826815. feature_fraction_stage2, val_score: 0.658928: 100%|##########| 3/3 [00:01<00:00, 1.97it/s] regularization_factors, val_score: 0.658928: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016489 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.578756 valid's binary_logloss: 0.666904 Early stopping, best iteration is: [60] train's binary_logloss: 0.606361 valid's binary_logloss: 0.662873 regularization_factors, val_score: 0.658928: 5%|5 | 1/20 [00:00<00:12, 1.52it/s][I 2020-09-27 04:42:53,400] Trial 40 finished with value: 0.6628730879769316 and parameters: {'lambda_l1': 0.00027860521966329625, 'lambda_l2': 0.23171527606664835}. Best is trial 40 with value: 0.6628730879769316. regularization_factors, val_score: 0.658928: 5%|5 | 1/20 [00:00<00:12, 1.52it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000377 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.57899 valid's binary_logloss: 0.664607 Early stopping, best iteration is: [60] train's binary_logloss: 0.607074 valid's binary_logloss: 0.658905 regularization_factors, val_score: 0.658905: 10%|# | 2/20 [00:01<00:11, 1.62it/s][I 2020-09-27 04:42:53,923] Trial 41 finished with value: 0.6589045464246605 and parameters: {'lambda_l1': 1.0562160571745345e-08, 'lambda_l2': 1.2335861617943539e-08}. Best is trial 41 with value: 0.6589045464246605. regularization_factors, val_score: 0.658905: 10%|# | 2/20 [00:01<00:11, 1.62it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000451 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.579401 valid's binary_logloss: 0.664539 Early stopping, best iteration is: [60] train's binary_logloss: 0.607081 valid's binary_logloss: 0.658927 regularization_factors, val_score: 0.658905: 15%|#5 | 3/20 [00:01<00:09, 1.70it/s][I 2020-09-27 04:42:54,440] Trial 42 finished with value: 0.6589271408352281 and parameters: {'lambda_l1': 1.4846795215967304e-08, 'lambda_l2': 2.1890701373620828e-08}. Best is trial 41 with value: 0.6589045464246605. regularization_factors, val_score: 0.658905: 15%|#5 | 3/20 [00:01<00:09, 1.70it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000474 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.579252 valid's binary_logloss: 0.665028 Early stopping, best iteration is: [60] train's binary_logloss: 0.607075 valid's binary_logloss: 0.658927 regularization_factors, val_score: 0.658905: 20%|## | 4/20 [00:02<00:09, 1.71it/s][I 2020-09-27 04:42:55,018] Trial 43 finished with value: 0.6589273467253626 and parameters: {'lambda_l1': 1.1433697445980986e-08, 'lambda_l2': 1.3023515499807511e-08}. Best is trial 41 with value: 0.6589045464246605. regularization_factors, val_score: 0.658905: 20%|## | 4/20 [00:02<00:09, 1.71it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000451 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.579216 valid's binary_logloss: 0.664431 Early stopping, best iteration is: [60] train's binary_logloss: 0.607079 valid's binary_logloss: 0.658949 regularization_factors, val_score: 0.658905: 25%|##5 | 5/20 [00:02<00:08, 1.75it/s][I 2020-09-27 04:42:55,554] Trial 44 finished with value: 0.6589487630970773 and parameters: {'lambda_l1': 1.0140056425681878e-08, 'lambda_l2': 1.2300058074804937e-08}. Best is trial 41 with value: 0.6589045464246605. regularization_factors, val_score: 0.658905: 25%|##5 | 5/20 [00:02<00:08, 1.75it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000479 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.578759 valid's binary_logloss: 0.665688 Early stopping, best iteration is: [60] train's binary_logloss: 0.607073 valid's binary_logloss: 0.658904 regularization_factors, val_score: 0.658904: 30%|### | 6/20 [00:03<00:07, 1.84it/s][I 2020-09-27 04:42:56,039] Trial 45 finished with value: 0.658904002539462 and parameters: {'lambda_l1': 1.0953495812808097e-08, 'lambda_l2': 1.828459432273254e-08}. Best is trial 45 with value: 0.658904002539462. regularization_factors, val_score: 0.658904: 30%|### | 6/20 [00:03<00:07, 1.84it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000378 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.579237 valid's binary_logloss: 0.665024 Early stopping, best iteration is: [60] train's binary_logloss: 0.60708 valid's binary_logloss: 0.658928 regularization_factors, val_score: 0.658904: 35%|###5 | 7/20 [00:04<00:08, 1.49it/s][I 2020-09-27 04:42:57,006] Trial 46 finished with value: 0.6589280313288517 and parameters: {'lambda_l1': 1.0619074780064361e-08, 'lambda_l2': 1.470079266120941e-08}. Best is trial 45 with value: 0.658904002539462. regularization_factors, val_score: 0.658904: 35%|###5 | 7/20 [00:04<00:08, 1.49it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000339 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.5794 valid's binary_logloss: 0.66454 Early stopping, best iteration is: [60] train's binary_logloss: 0.607078 valid's binary_logloss: 0.658904 regularization_factors, val_score: 0.658904: 40%|#### | 8/20 [00:04<00:07, 1.63it/s][I 2020-09-27 04:42:57,486] Trial 47 finished with value: 0.6589040025411772 and parameters: {'lambda_l1': 1.3254321926524086e-08, 'lambda_l2': 2.193885866761881e-08}. Best is trial 45 with value: 0.658904002539462. regularization_factors, val_score: 0.658904: 40%|#### | 8/20 [00:04<00:07, 1.63it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000424 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.579223 valid's binary_logloss: 0.664396 Early stopping, best iteration is: [60] train's binary_logloss: 0.607078 valid's binary_logloss: 0.658927 regularization_factors, val_score: 0.658904: 45%|####5 | 9/20 [00:05<00:06, 1.76it/s][I 2020-09-27 04:42:57,951] Trial 48 finished with value: 0.6589267944230356 and parameters: {'lambda_l1': 2.4619075323767072e-08, 'lambda_l2': 2.017164260358036e-08}. Best is trial 45 with value: 0.658904002539462. regularization_factors, val_score: 0.658904: 45%|####5 | 9/20 [00:05<00:06, 1.76it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000536 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.57921 valid's binary_logloss: 0.664407 Early stopping, best iteration is: [60] train's binary_logloss: 0.607077 valid's binary_logloss: 0.658925 regularization_factors, val_score: 0.658904: 50%|##### | 10/20 [00:05<00:05, 1.81it/s][I 2020-09-27 04:42:58,460] Trial 49 finished with value: 0.6589253929659314 and parameters: {'lambda_l1': 1.2653815920250245e-08, 'lambda_l2': 2.3139792938926494e-08}. Best is trial 45 with value: 0.658904002539462. regularization_factors, val_score: 0.658904: 50%|##### | 10/20 [00:05<00:05, 1.81it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004484 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.579211 valid's binary_logloss: 0.664425 Early stopping, best iteration is: [60] train's binary_logloss: 0.607081 valid's binary_logloss: 0.658927 regularization_factors, val_score: 0.658904: 55%|#####5 | 11/20 [00:06<00:04, 1.90it/s][I 2020-09-27 04:42:58,929] Trial 50 finished with value: 0.6589271408099914 and parameters: {'lambda_l1': 1.3776160562135245e-08, 'lambda_l2': 2.6228878428660015e-07}. Best is trial 45 with value: 0.658904002539462. regularization_factors, val_score: 0.658904: 55%|#####5 | 11/20 [00:06<00:04, 1.90it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000373 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.57921 valid's binary_logloss: 0.664409 Early stopping, best iteration is: [60] train's binary_logloss: 0.607081 valid's binary_logloss: 0.658927 regularization_factors, val_score: 0.658904: 60%|###### | 12/20 [00:06<00:04, 1.94it/s][I 2020-09-27 04:42:59,423] Trial 51 finished with value: 0.658927140833919 and parameters: {'lambda_l1': 1.664666646319637e-08, 'lambda_l2': 4.250351022086274e-08}. Best is trial 45 with value: 0.658904002539462. regularization_factors, val_score: 0.658904: 60%|###### | 12/20 [00:06<00:04, 1.94it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000390 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.57924 valid's binary_logloss: 0.665044 Early stopping, best iteration is: [60] train's binary_logloss: 0.607082 valid's binary_logloss: 0.65895 regularization_factors, val_score: 0.658904: 65%|######5 | 13/20 [00:07<00:03, 1.94it/s][I 2020-09-27 04:42:59,937] Trial 52 finished with value: 0.6589497680473425 and parameters: {'lambda_l1': 1.439041833170667e-08, 'lambda_l2': 3.7007408240165166e-07}. Best is trial 45 with value: 0.658904002539462. regularization_factors, val_score: 0.658904: 65%|######5 | 13/20 [00:07<00:03, 1.94it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000634 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.579255 valid's binary_logloss: 0.665048 Early stopping, best iteration is: [60] train's binary_logloss: 0.607079 valid's binary_logloss: 0.658949 regularization_factors, val_score: 0.658904: 70%|####### | 14/20 [00:08<00:04, 1.47it/s][I 2020-09-27 04:43:01,006] Trial 53 finished with value: 0.6589494209805504 and parameters: {'lambda_l1': 1.7572183872374385e-07, 'lambda_l2': 9.82651046880018e-07}. Best is trial 45 with value: 0.658904002539462. regularization_factors, val_score: 0.658904: 70%|####### | 14/20 [00:08<00:04, 1.47it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000481 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.579214 valid's binary_logloss: 0.664386 Early stopping, best iteration is: [60] train's binary_logloss: 0.607075 valid's binary_logloss: 0.658904 regularization_factors, val_score: 0.658904: 75%|#######5 | 15/20 [00:08<00:03, 1.60it/s][I 2020-09-27 04:43:01,494] Trial 54 finished with value: 0.6589036559607917 and parameters: {'lambda_l1': 8.538730576389241e-07, 'lambda_l2': 1.1721978821541268e-08}. Best is trial 54 with value: 0.6589036559607917. regularization_factors, val_score: 0.658904: 75%|#######5 | 15/20 [00:08<00:03, 1.60it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000414 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.579216 valid's binary_logloss: 0.664476 Early stopping, best iteration is: [60] train's binary_logloss: 0.607081 valid's binary_logloss: 0.658927 regularization_factors, val_score: 0.658904: 80%|######## | 16/20 [00:09<00:02, 1.72it/s][I 2020-09-27 04:43:01,976] Trial 55 finished with value: 0.6589271405882069 and parameters: {'lambda_l1': 2.1323929878136295e-06, 'lambda_l2': 1.2914444598331435e-08}. Best is trial 54 with value: 0.6589036559607917. regularization_factors, val_score: 0.658904: 80%|######## | 16/20 [00:09<00:02, 1.72it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004379 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.579258 valid's binary_logloss: 0.665048 Early stopping, best iteration is: [60] train's binary_logloss: 0.607079 valid's binary_logloss: 0.658949 regularization_factors, val_score: 0.658904: 85%|########5 | 17/20 [00:09<00:01, 1.82it/s][I 2020-09-27 04:43:02,455] Trial 56 finished with value: 0.6589494210443969 and parameters: {'lambda_l1': 7.412222130813083e-07, 'lambda_l2': 1.1296738167689051e-08}. Best is trial 54 with value: 0.6589036559607917. regularization_factors, val_score: 0.658904: 85%|########5 | 17/20 [00:09<00:01, 1.82it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000368 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.579256 valid's binary_logloss: 0.665048 Early stopping, best iteration is: [60] train's binary_logloss: 0.607077 valid's binary_logloss: 0.658949 regularization_factors, val_score: 0.658904: 90%|######### | 18/20 [00:10<00:01, 1.87it/s][I 2020-09-27 04:43:02,951] Trial 57 finished with value: 0.6589493135745349 and parameters: {'lambda_l1': 2.9906109268867005e-07, 'lambda_l2': 1.0924154779547353e-05}. Best is trial 54 with value: 0.6589036559607917. regularization_factors, val_score: 0.658904: 90%|######### | 18/20 [00:10<00:01, 1.87it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000413 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.579254 valid's binary_logloss: 0.665027 Early stopping, best iteration is: [60] train's binary_logloss: 0.607078 valid's binary_logloss: 0.658927 regularization_factors, val_score: 0.658904: 95%|#########5| 19/20 [00:10<00:00, 1.95it/s][I 2020-09-27 04:43:03,413] Trial 58 finished with value: 0.658926859627446 and parameters: {'lambda_l1': 1.208558071943979e-07, 'lambda_l2': 1.4153597871067393e-07}. Best is trial 54 with value: 0.6589036559607917. regularization_factors, val_score: 0.658904: 95%|#########5| 19/20 [00:10<00:00, 1.95it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004727 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.579258 valid's binary_logloss: 0.665048 Early stopping, best iteration is: [60] train's binary_logloss: 0.607079 valid's binary_logloss: 0.658949 regularization_factors, val_score: 0.658904: 100%|##########| 20/20 [00:11<00:00, 1.94it/s][I 2020-09-27 04:43:03,937] Trial 59 finished with value: 0.6589494193426407 and parameters: {'lambda_l1': 1.386319739479547e-05, 'lambda_l2': 1.070237025489994e-08}. Best is trial 54 with value: 0.6589036559607917. regularization_factors, val_score: 0.658904: 100%|##########| 20/20 [00:11<00:00, 1.78it/s] min_data_in_leaf, val_score: 0.658904: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000392 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.576809 valid's binary_logloss: 0.668517 Early stopping, best iteration is: [60] train's binary_logloss: 0.605306 valid's binary_logloss: 0.663449 min_data_in_leaf, val_score: 0.658904: 20%|## | 1/5 [00:00<00:03, 1.01it/s][I 2020-09-27 04:43:04,955] Trial 60 finished with value: 0.6634487376945314 and parameters: {'min_child_samples': 5}. Best is trial 60 with value: 0.6634487376945314. min_data_in_leaf, val_score: 0.658904: 20%|## | 1/5 [00:01<00:03, 1.01it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000379 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.579781 valid's binary_logloss: 0.664717 Early stopping, best iteration is: [60] train's binary_logloss: 0.607283 valid's binary_logloss: 0.661178 min_data_in_leaf, val_score: 0.658904: 40%|#### | 2/5 [00:01<00:02, 1.17it/s][I 2020-09-27 04:43:05,479] Trial 61 finished with value: 0.6611778602130459 and parameters: {'min_child_samples': 25}. Best is trial 61 with value: 0.6611778602130459. min_data_in_leaf, val_score: 0.658904: 40%|#### | 2/5 [00:01<00:02, 1.17it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004590 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.587755 valid's binary_logloss: 0.665012 Early stopping, best iteration is: [66] train's binary_logloss: 0.607178 valid's binary_logloss: 0.662389 min_data_in_leaf, val_score: 0.658904: 60%|###### | 3/5 [00:02<00:01, 1.34it/s][I 2020-09-27 04:43:05,977] Trial 62 finished with value: 0.6623889382406033 and parameters: {'min_child_samples': 100}. Best is trial 61 with value: 0.6611778602130459. min_data_in_leaf, val_score: 0.658904: 60%|###### | 3/5 [00:02<00:01, 1.34it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000482 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.577819 valid's binary_logloss: 0.667336 Early stopping, best iteration is: [60] train's binary_logloss: 0.605667 valid's binary_logloss: 0.663345 min_data_in_leaf, val_score: 0.658904: 80%|######## | 4/5 [00:02<00:00, 1.50it/s][I 2020-09-27 04:43:06,461] Trial 63 finished with value: 0.6633454558808694 and parameters: {'min_child_samples': 10}. Best is trial 61 with value: 0.6611778602130459. min_data_in_leaf, val_score: 0.658904: 80%|######## | 4/5 [00:02<00:00, 1.50it/s][LightGBM] [Info] Number of positive: 13155, number of negative: 12844 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000449 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.505981 -> initscore=0.023925 [LightGBM] [Info] Start training from score 0.023925 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.583153 valid's binary_logloss: 0.66795 Early stopping, best iteration is: [59] train's binary_logloss: 0.610418 valid's binary_logloss: 0.663074 min_data_in_leaf, val_score: 0.658904: 100%|##########| 5/5 [00:02<00:00, 1.63it/s][I 2020-09-27 04:43:06,953] Trial 64 finished with value: 0.6630735000484437 and parameters: {'min_child_samples': 50}. Best is trial 61 with value: 0.6611778602130459. min_data_in_leaf, val_score: 0.658904: 100%|##########| 5/5 [00:02<00:00, 1.67it/s] Fold : 5 [I 2020-09-27 04:43:06,989] A new study created in memory with name: no-name-b151d15a-5a47-42dc-ace0-4fe96bb3c92c feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004614 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.577114 valid's binary_logloss: 0.654088 Early stopping, best iteration is: [60] train's binary_logloss: 0.604064 valid's binary_logloss: 0.652612 feature_fraction, val_score: 0.652612: 14%|#4 | 1/7 [00:00<00:02, 2.09it/s][I 2020-09-27 04:43:07,480] Trial 0 finished with value: 0.6526118300374438 and parameters: {'feature_fraction': 0.6}. Best is trial 0 with value: 0.6526118300374438. feature_fraction, val_score: 0.652612: 14%|#4 | 1/7 [00:00<00:02, 2.09it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001021 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.570555 valid's binary_logloss: 0.652172 [200] train's binary_logloss: 0.514477 valid's binary_logloss: 0.655545 Early stopping, best iteration is: [114] train's binary_logloss: 0.561489 valid's binary_logloss: 0.651258 feature_fraction, val_score: 0.651258: 29%|##8 | 2/7 [00:01<00:03, 1.52it/s][I 2020-09-27 04:43:08,551] Trial 1 finished with value: 0.6512579255125732 and parameters: {'feature_fraction': 1.0}. Best is trial 1 with value: 0.6512579255125732. feature_fraction, val_score: 0.651258: 29%|##8 | 2/7 [00:01<00:03, 1.52it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004468 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.583108 valid's binary_logloss: 0.653379 Early stopping, best iteration is: [83] train's binary_logloss: 0.593455 valid's binary_logloss: 0.652673 feature_fraction, val_score: 0.651258: 43%|####2 | 3/7 [00:01<00:02, 1.69it/s][I 2020-09-27 04:43:08,993] Trial 2 finished with value: 0.652673016771765 and parameters: {'feature_fraction': 0.4}. Best is trial 1 with value: 0.6512579255125732. feature_fraction, val_score: 0.651258: 43%|####2 | 3/7 [00:01<00:02, 1.69it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000894 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.573161 valid's binary_logloss: 0.653887 [200] train's binary_logloss: 0.518397 valid's binary_logloss: 0.656355 Early stopping, best iteration is: [117] train's binary_logloss: 0.562873 valid's binary_logloss: 0.652985 feature_fraction, val_score: 0.651258: 57%|#####7 | 4/7 [00:02<00:01, 1.66it/s][I 2020-09-27 04:43:09,624] Trial 3 finished with value: 0.6529847756787174 and parameters: {'feature_fraction': 0.8}. Best is trial 1 with value: 0.6512579255125732. feature_fraction, val_score: 0.651258: 57%|#####7 | 4/7 [00:02<00:01, 1.66it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000779 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.574837 valid's binary_logloss: 0.653256 Early stopping, best iteration is: [78] train's binary_logloss: 0.589091 valid's binary_logloss: 0.651848 feature_fraction, val_score: 0.651258: 71%|#######1 | 5/7 [00:03<00:01, 1.75it/s][I 2020-09-27 04:43:10,125] Trial 4 finished with value: 0.6518483772036818 and parameters: {'feature_fraction': 0.7}. Best is trial 1 with value: 0.6512579255125732. feature_fraction, val_score: 0.651258: 71%|#######1 | 5/7 [00:03<00:01, 1.75it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000916 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.572031 valid's binary_logloss: 0.652234 Early stopping, best iteration is: [76] train's binary_logloss: 0.588368 valid's binary_logloss: 0.651017 feature_fraction, val_score: 0.651017: 86%|########5 | 6/7 [00:03<00:00, 1.79it/s][I 2020-09-27 04:43:10,646] Trial 5 finished with value: 0.6510174434829193 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 5 with value: 0.6510174434829193. feature_fraction, val_score: 0.651017: 86%|########5 | 6/7 [00:03<00:00, 1.79it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000533 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.578051 valid's binary_logloss: 0.652005 [200] train's binary_logloss: 0.525344 valid's binary_logloss: 0.654096 Early stopping, best iteration is: [137] train's binary_logloss: 0.557437 valid's binary_logloss: 0.651283 feature_fraction, val_score: 0.651017: 100%|##########| 7/7 [00:04<00:00, 1.76it/s][I 2020-09-27 04:43:11,237] Trial 6 finished with value: 0.6512826266964433 and parameters: {'feature_fraction': 0.5}. Best is trial 5 with value: 0.6510174434829193. feature_fraction, val_score: 0.651017: 100%|##########| 7/7 [00:04<00:00, 1.65it/s] num_leaves, val_score: 0.651017: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000858 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.519602 valid's binary_logloss: 0.657482 Early stopping, best iteration is: [48] train's binary_logloss: 0.578782 valid's binary_logloss: 0.655742 num_leaves, val_score: 0.651017: 5%|5 | 1/20 [00:01<00:23, 1.23s/it][I 2020-09-27 04:43:12,484] Trial 7 finished with value: 0.655741889936831 and parameters: {'num_leaves': 55}. Best is trial 7 with value: 0.655741889936831. num_leaves, val_score: 0.651017: 5%|5 | 1/20 [00:01<00:23, 1.23s/it][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000474 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.352805 valid's binary_logloss: 0.663625 Early stopping, best iteration is: [40] train's binary_logloss: 0.492845 valid's binary_logloss: 0.6534 num_leaves, val_score: 0.651017: 10%|# | 2/20 [00:02<00:21, 1.18s/it][I 2020-09-27 04:43:13,542] Trial 8 finished with value: 0.653399633240566 and parameters: {'num_leaves': 164}. Best is trial 8 with value: 0.653399633240566. num_leaves, val_score: 0.651017: 10%|# | 2/20 [00:02<00:21, 1.18s/it][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005116 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.645865 valid's binary_logloss: 0.655865 [200] train's binary_logloss: 0.633609 valid's binary_logloss: 0.652415 [300] train's binary_logloss: 0.624725 valid's binary_logloss: 0.651323 Early stopping, best iteration is: [292] train's binary_logloss: 0.625397 valid's binary_logloss: 0.65122 num_leaves, val_score: 0.651017: 15%|#5 | 3/20 [00:02<00:17, 1.02s/it][I 2020-09-27 04:43:14,213] Trial 9 finished with value: 0.6512203104712043 and parameters: {'num_leaves': 5}. Best is trial 9 with value: 0.6512203104712043. num_leaves, val_score: 0.651017: 15%|#5 | 3/20 [00:02<00:17, 1.02s/it][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000891 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.297872 valid's binary_logloss: 0.671789 Early stopping, best iteration is: [27] train's binary_logloss: 0.510065 valid's binary_logloss: 0.658243 num_leaves, val_score: 0.651017: 20%|## | 4/20 [00:04<00:20, 1.30s/it][I 2020-09-27 04:43:16,157] Trial 10 finished with value: 0.6582425249764476 and parameters: {'num_leaves': 215}. Best is trial 9 with value: 0.6512203104712043. num_leaves, val_score: 0.651017: 20%|## | 4/20 [00:04<00:20, 1.30s/it][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000907 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.669067 valid's binary_logloss: 0.671099 [200] train's binary_logloss: 0.659436 valid's binary_logloss: 0.661559 [300] train's binary_logloss: 0.654191 valid's binary_logloss: 0.656836 [400] train's binary_logloss: 0.651007 valid's binary_logloss: 0.65419 [500] train's binary_logloss: 0.64896 valid's binary_logloss: 0.652841 [600] train's binary_logloss: 0.647587 valid's binary_logloss: 0.652048 [700] train's binary_logloss: 0.646612 valid's binary_logloss: 0.651613 [800] train's binary_logloss: 0.645876 valid's binary_logloss: 0.651451 [900] train's binary_logloss: 0.645296 valid's binary_logloss: 0.651421 Early stopping, best iteration is: [887] train's binary_logloss: 0.645365 valid's binary_logloss: 0.651313 num_leaves, val_score: 0.651017: 25%|##5 | 5/20 [00:06<00:19, 1.29s/it][I 2020-09-27 04:43:17,415] Trial 11 finished with value: 0.6513127631312726 and parameters: {'num_leaves': 2}. Best is trial 9 with value: 0.6512203104712043. num_leaves, val_score: 0.651017: 25%|##5 | 5/20 [00:06<00:19, 1.29s/it][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001145 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.487358 valid's binary_logloss: 0.659662 Early stopping, best iteration is: [35] train's binary_logloss: 0.581991 valid's binary_logloss: 0.657768 num_leaves, val_score: 0.651017: 30%|### | 6/20 [00:06<00:15, 1.10s/it][I 2020-09-27 04:43:18,063] Trial 12 finished with value: 0.6577682570812166 and parameters: {'num_leaves': 73}. Best is trial 9 with value: 0.6512203104712043. num_leaves, val_score: 0.651017: 30%|### | 6/20 [00:06<00:15, 1.10s/it][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001036 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.607358 valid's binary_logloss: 0.650129 [200] train's binary_logloss: 0.572939 valid's binary_logloss: 0.650833 Early stopping, best iteration is: [137] train's binary_logloss: 0.593957 valid's binary_logloss: 0.649769 num_leaves, val_score: 0.649769: 35%|###5 | 7/20 [00:07<00:12, 1.07it/s][I 2020-09-27 04:43:18,630] Trial 13 finished with value: 0.649768864866138 and parameters: {'num_leaves': 17}. Best is trial 13 with value: 0.649768864866138. num_leaves, val_score: 0.649769: 35%|###5 | 7/20 [00:07<00:12, 1.07it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001092 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.41967 valid's binary_logloss: 0.663995 Early stopping, best iteration is: [33] train's binary_logloss: 0.552391 valid's binary_logloss: 0.657367 num_leaves, val_score: 0.649769: 40%|#### | 8/20 [00:08<00:12, 1.07s/it][I 2020-09-27 04:43:20,007] Trial 14 finished with value: 0.6573672765256382 and parameters: {'num_leaves': 114}. Best is trial 13 with value: 0.649768864866138. num_leaves, val_score: 0.649769: 40%|#### | 8/20 [00:08<00:12, 1.07s/it][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000964 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.262043 valid's binary_logloss: 0.668022 Early stopping, best iteration is: [31] train's binary_logloss: 0.470728 valid's binary_logloss: 0.656929 num_leaves, val_score: 0.649769: 45%|####5 | 9/20 [00:10<00:13, 1.26s/it][I 2020-09-27 04:43:21,707] Trial 15 finished with value: 0.6569287641382702 and parameters: {'num_leaves': 256}. Best is trial 13 with value: 0.649768864866138. num_leaves, val_score: 0.649769: 45%|####5 | 9/20 [00:10<00:13, 1.26s/it][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000921 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.534752 valid's binary_logloss: 0.65496 Early stopping, best iteration is: [63] train's binary_logloss: 0.569855 valid's binary_logloss: 0.652951 num_leaves, val_score: 0.649769: 50%|##### | 10/20 [00:11<00:10, 1.07s/it][I 2020-09-27 04:43:22,329] Trial 16 finished with value: 0.6529514815127009 and parameters: {'num_leaves': 48}. Best is trial 13 with value: 0.649768864866138. num_leaves, val_score: 0.649769: 50%|##### | 10/20 [00:11<00:10, 1.07s/it][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004780 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.382444 valid's binary_logloss: 0.663277 Early stopping, best iteration is: [36] train's binary_logloss: 0.523542 valid's binary_logloss: 0.654707 num_leaves, val_score: 0.649769: 55%|#####5 | 11/20 [00:12<00:10, 1.12s/it][I 2020-09-27 04:43:23,589] Trial 17 finished with value: 0.6547072669652428 and parameters: {'num_leaves': 140}. Best is trial 13 with value: 0.649768864866138. num_leaves, val_score: 0.649769: 55%|#####5 | 11/20 [00:12<00:10, 1.12s/it][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.003413 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.448357 valid's binary_logloss: 0.661818 Early stopping, best iteration is: [46] train's binary_logloss: 0.538243 valid's binary_logloss: 0.657096 num_leaves, val_score: 0.649769: 60%|###### | 12/20 [00:13<00:08, 1.08s/it][I 2020-09-27 04:43:24,567] Trial 18 finished with value: 0.6570959414584372 and parameters: {'num_leaves': 95}. Best is trial 13 with value: 0.649768864866138. num_leaves, val_score: 0.649769: 60%|###### | 12/20 [00:13<00:08, 1.08s/it][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010468 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.589787 valid's binary_logloss: 0.652218 Early stopping, best iteration is: [97] train's binary_logloss: 0.591353 valid's binary_logloss: 0.652064 num_leaves, val_score: 0.649769: 65%|######5 | 13/20 [00:13<00:06, 1.09it/s][I 2020-09-27 04:43:25,099] Trial 19 finished with value: 0.6520641906266819 and parameters: {'num_leaves': 24}. Best is trial 13 with value: 0.649768864866138. num_leaves, val_score: 0.649769: 65%|######5 | 13/20 [00:13<00:06, 1.09it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000894 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.332931 valid's binary_logloss: 0.669943 Early stopping, best iteration is: [24] train's binary_logloss: 0.541056 valid's binary_logloss: 0.658153 num_leaves, val_score: 0.649769: 70%|####### | 14/20 [00:14<00:05, 1.02it/s][I 2020-09-27 04:43:26,213] Trial 20 finished with value: 0.6581527972889681 and parameters: {'num_leaves': 181}. Best is trial 13 with value: 0.649768864866138. num_leaves, val_score: 0.649769: 70%|####### | 14/20 [00:14<00:05, 1.02it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010464 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.645865 valid's binary_logloss: 0.655865 [200] train's binary_logloss: 0.633609 valid's binary_logloss: 0.652415 [300] train's binary_logloss: 0.624725 valid's binary_logloss: 0.651323 Early stopping, best iteration is: [292] train's binary_logloss: 0.625397 valid's binary_logloss: 0.65122 num_leaves, val_score: 0.649769: 75%|#######5 | 15/20 [00:15<00:04, 1.13it/s][I 2020-09-27 04:43:26,886] Trial 21 finished with value: 0.6512203104712043 and parameters: {'num_leaves': 5}. Best is trial 13 with value: 0.649768864866138. num_leaves, val_score: 0.649769: 75%|#######5 | 15/20 [00:15<00:04, 1.13it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000872 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.579189 valid's binary_logloss: 0.654451 [200] train's binary_logloss: 0.527779 valid's binary_logloss: 0.653698 Early stopping, best iteration is: [151] train's binary_logloss: 0.551963 valid's binary_logloss: 0.653462 num_leaves, val_score: 0.649769: 80%|######## | 16/20 [00:16<00:04, 1.00s/it][I 2020-09-27 04:43:28,167] Trial 22 finished with value: 0.6534616704278683 and parameters: {'num_leaves': 28}. Best is trial 13 with value: 0.649768864866138. num_leaves, val_score: 0.649769: 80%|######## | 16/20 [00:16<00:04, 1.00s/it][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011614 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.472661 valid's binary_logloss: 0.658505 Early stopping, best iteration is: [28] train's binary_logloss: 0.589272 valid's binary_logloss: 0.658288 num_leaves, val_score: 0.649769: 85%|########5 | 17/20 [00:17<00:02, 1.13it/s][I 2020-09-27 04:43:28,784] Trial 23 finished with value: 0.658288148145066 and parameters: {'num_leaves': 81}. Best is trial 13 with value: 0.649768864866138. num_leaves, val_score: 0.649769: 85%|########5 | 17/20 [00:17<00:02, 1.13it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000925 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.645865 valid's binary_logloss: 0.655865 [200] train's binary_logloss: 0.633609 valid's binary_logloss: 0.652415 [300] train's binary_logloss: 0.624725 valid's binary_logloss: 0.651323 Early stopping, best iteration is: [292] train's binary_logloss: 0.625397 valid's binary_logloss: 0.65122 num_leaves, val_score: 0.649769: 90%|######### | 18/20 [00:18<00:01, 1.23it/s][I 2020-09-27 04:43:29,433] Trial 24 finished with value: 0.6512203104712043 and parameters: {'num_leaves': 5}. Best is trial 13 with value: 0.649768864866138. num_leaves, val_score: 0.649769: 90%|######### | 18/20 [00:18<00:01, 1.23it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004987 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.565987 valid's binary_logloss: 0.655697 Early stopping, best iteration is: [59] train's binary_logloss: 0.596142 valid's binary_logloss: 0.65418 num_leaves, val_score: 0.649769: 95%|#########5| 19/20 [00:18<00:00, 1.34it/s][I 2020-09-27 04:43:30,009] Trial 25 finished with value: 0.6541801455639907 and parameters: {'num_leaves': 34}. Best is trial 13 with value: 0.649768864866138. num_leaves, val_score: 0.649769: 95%|#########5| 19/20 [00:18<00:00, 1.34it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010666 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.512496 valid's binary_logloss: 0.657066 Early stopping, best iteration is: [78] train's binary_logloss: 0.536548 valid's binary_logloss: 0.654935 num_leaves, val_score: 0.649769: 100%|##########| 20/20 [00:19<00:00, 1.38it/s][I 2020-09-27 04:43:30,691] Trial 26 finished with value: 0.6549352289587036 and parameters: {'num_leaves': 59}. Best is trial 13 with value: 0.649768864866138. num_leaves, val_score: 0.649769: 100%|##########| 20/20 [00:19<00:00, 1.03it/s] bagging, val_score: 0.649769: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000913 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.612583 valid's binary_logloss: 0.656379 Early stopping, best iteration is: [86] train's binary_logloss: 0.617678 valid's binary_logloss: 0.655144 bagging, val_score: 0.649769: 10%|# | 1/10 [00:00<00:07, 1.17it/s][I 2020-09-27 04:43:31,563] Trial 27 finished with value: 0.6551437821744738 and parameters: {'bagging_fraction': 0.40320944016041527, 'bagging_freq': 7}. Best is trial 27 with value: 0.6551437821744738. bagging, val_score: 0.649769: 10%|# | 1/10 [00:00<00:07, 1.17it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.014340 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.607438 valid's binary_logloss: 0.653472 [200] train's binary_logloss: 0.573071 valid's binary_logloss: 0.653346 Early stopping, best iteration is: [139] train's binary_logloss: 0.593171 valid's binary_logloss: 0.652081 bagging, val_score: 0.649769: 20%|## | 2/10 [00:01<00:06, 1.21it/s][I 2020-09-27 04:43:32,320] Trial 28 finished with value: 0.6520811290472097 and parameters: {'bagging_fraction': 0.9561159020668242, 'bagging_freq': 1}. Best is trial 28 with value: 0.6520811290472097. bagging, val_score: 0.649769: 20%|## | 2/10 [00:01<00:06, 1.21it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001122 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.610187 valid's binary_logloss: 0.652923 Early stopping, best iteration is: [81] train's binary_logloss: 0.617563 valid's binary_logloss: 0.651804 bagging, val_score: 0.649769: 30%|### | 3/10 [00:02<00:05, 1.39it/s][I 2020-09-27 04:43:32,796] Trial 29 finished with value: 0.6518044921241909 and parameters: {'bagging_fraction': 0.5465375900437621, 'bagging_freq': 4}. Best is trial 29 with value: 0.6518044921241909. bagging, val_score: 0.649769: 30%|### | 3/10 [00:02<00:05, 1.39it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002328 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.607479 valid's binary_logloss: 0.651182 [200] train's binary_logloss: 0.571824 valid's binary_logloss: 0.653082 Early stopping, best iteration is: [118] train's binary_logloss: 0.600602 valid's binary_logloss: 0.650679 bagging, val_score: 0.649769: 40%|#### | 4/10 [00:02<00:04, 1.46it/s][I 2020-09-27 04:43:33,403] Trial 30 finished with value: 0.6506791175850941 and parameters: {'bagging_fraction': 0.9711225430872865, 'bagging_freq': 7}. Best is trial 30 with value: 0.6506791175850941. bagging, val_score: 0.649769: 40%|#### | 4/10 [00:02<00:04, 1.46it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005891 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.607508 valid's binary_logloss: 0.652005 [200] train's binary_logloss: 0.572706 valid's binary_logloss: 0.650621 Early stopping, best iteration is: [129] train's binary_logloss: 0.596726 valid's binary_logloss: 0.650382 bagging, val_score: 0.649769: 50%|##### | 5/10 [00:03<00:03, 1.48it/s][I 2020-09-27 04:43:34,052] Trial 31 finished with value: 0.6503824259794613 and parameters: {'bagging_fraction': 0.9985128973254092, 'bagging_freq': 7}. Best is trial 31 with value: 0.6503824259794613. bagging, val_score: 0.649769: 50%|##### | 5/10 [00:03<00:03, 1.48it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005483 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.60784 valid's binary_logloss: 0.652715 [200] train's binary_logloss: 0.572748 valid's binary_logloss: 0.653478 Early stopping, best iteration is: [124] train's binary_logloss: 0.598892 valid's binary_logloss: 0.651677 bagging, val_score: 0.649769: 60%|###### | 6/10 [00:04<00:02, 1.48it/s][I 2020-09-27 04:43:34,731] Trial 32 finished with value: 0.6516771078927124 and parameters: {'bagging_fraction': 0.9937517875557063, 'bagging_freq': 7}. Best is trial 31 with value: 0.6503824259794613. bagging, val_score: 0.649769: 60%|###### | 6/10 [00:04<00:02, 1.48it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.003559 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.607577 valid's binary_logloss: 0.652062 [200] train's binary_logloss: 0.572523 valid's binary_logloss: 0.651285 [300] train's binary_logloss: 0.541985 valid's binary_logloss: 0.651463 Early stopping, best iteration is: [260] train's binary_logloss: 0.554233 valid's binary_logloss: 0.65052 bagging, val_score: 0.649769: 70%|####### | 7/10 [00:05<00:02, 1.13it/s][I 2020-09-27 04:43:36,093] Trial 33 finished with value: 0.6505203576283998 and parameters: {'bagging_fraction': 0.8468420031688184, 'bagging_freq': 6}. Best is trial 31 with value: 0.6503824259794613. bagging, val_score: 0.649769: 70%|####### | 7/10 [00:05<00:02, 1.13it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001043 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.60809 valid's binary_logloss: 0.653721 Early stopping, best iteration is: [83] train's binary_logloss: 0.614817 valid's binary_logloss: 0.653287 bagging, val_score: 0.649769: 80%|######## | 8/10 [00:05<00:01, 1.30it/s][I 2020-09-27 04:43:36,593] Trial 34 finished with value: 0.6532867425106659 and parameters: {'bagging_fraction': 0.8464729466696903, 'bagging_freq': 7}. Best is trial 31 with value: 0.6503824259794613. bagging, val_score: 0.649769: 80%|######## | 8/10 [00:05<00:01, 1.30it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000488 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.608096 valid's binary_logloss: 0.652476 [200] train's binary_logloss: 0.572069 valid's binary_logloss: 0.654848 Early stopping, best iteration is: [105] train's binary_logloss: 0.606068 valid's binary_logloss: 0.652367 bagging, val_score: 0.649769: 90%|######### | 9/10 [00:06<00:00, 1.41it/s][I 2020-09-27 04:43:37,161] Trial 35 finished with value: 0.6523668340051799 and parameters: {'bagging_fraction': 0.830591192364971, 'bagging_freq': 5}. Best is trial 31 with value: 0.6503824259794613. bagging, val_score: 0.649769: 90%|######### | 9/10 [00:06<00:00, 1.41it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000998 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.607246 valid's binary_logloss: 0.652045 Early stopping, best iteration is: [95] train's binary_logloss: 0.609255 valid's binary_logloss: 0.651848 bagging, val_score: 0.649769: 100%|##########| 10/10 [00:07<00:00, 1.50it/s][I 2020-09-27 04:43:37,726] Trial 36 finished with value: 0.6518478119160271 and parameters: {'bagging_fraction': 0.9992112947235333, 'bagging_freq': 6}. Best is trial 31 with value: 0.6503824259794613. bagging, val_score: 0.649769: 100%|##########| 10/10 [00:07<00:00, 1.42it/s] feature_fraction_stage2, val_score: 0.649769: 0%| | 0/6 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000951 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.606705 valid's binary_logloss: 0.652233 [200] train's binary_logloss: 0.572023 valid's binary_logloss: 0.650662 [300] train's binary_logloss: 0.542509 valid's binary_logloss: 0.653346 Early stopping, best iteration is: [200] train's binary_logloss: 0.572023 valid's binary_logloss: 0.650662 feature_fraction_stage2, val_score: 0.649769: 17%|#6 | 1/6 [00:00<00:03, 1.42it/s][I 2020-09-27 04:43:38,443] Trial 37 finished with value: 0.6506622068727096 and parameters: {'feature_fraction': 0.9799999999999999}. Best is trial 37 with value: 0.6506622068727096. feature_fraction_stage2, val_score: 0.649769: 17%|#6 | 1/6 [00:00<00:03, 1.42it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000943 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.607972 valid's binary_logloss: 0.651443 [200] train's binary_logloss: 0.573729 valid's binary_logloss: 0.651681 Early stopping, best iteration is: [174] train's binary_logloss: 0.58195 valid's binary_logloss: 0.650991 feature_fraction_stage2, val_score: 0.649769: 33%|###3 | 2/6 [00:01<00:03, 1.17it/s][I 2020-09-27 04:43:39,648] Trial 38 finished with value: 0.6509908340902059 and parameters: {'feature_fraction': 0.852}. Best is trial 37 with value: 0.6506622068727096. feature_fraction_stage2, val_score: 0.649769: 33%|###3 | 2/6 [00:01<00:03, 1.17it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000929 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.606705 valid's binary_logloss: 0.652233 [200] train's binary_logloss: 0.572023 valid's binary_logloss: 0.650662 [300] train's binary_logloss: 0.542509 valid's binary_logloss: 0.653346 Early stopping, best iteration is: [200] train's binary_logloss: 0.572023 valid's binary_logloss: 0.650662 feature_fraction_stage2, val_score: 0.649769: 50%|##### | 3/6 [00:02<00:02, 1.22it/s][I 2020-09-27 04:43:40,396] Trial 39 finished with value: 0.6506622068727096 and parameters: {'feature_fraction': 0.948}. Best is trial 37 with value: 0.6506622068727096. feature_fraction_stage2, val_score: 0.649769: 50%|##### | 3/6 [00:02<00:02, 1.22it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000871 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.60867 valid's binary_logloss: 0.651571 Early stopping, best iteration is: [95] train's binary_logloss: 0.610608 valid's binary_logloss: 0.651188 feature_fraction_stage2, val_score: 0.649769: 67%|######6 | 4/6 [00:03<00:01, 1.37it/s][I 2020-09-27 04:43:40,904] Trial 40 finished with value: 0.6511879268615787 and parameters: {'feature_fraction': 0.82}. Best is trial 37 with value: 0.6506622068727096. feature_fraction_stage2, val_score: 0.649769: 67%|######6 | 4/6 [00:03<00:01, 1.37it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005317 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.60735 valid's binary_logloss: 0.652555 Early stopping, best iteration is: [72] train's binary_logloss: 0.618944 valid's binary_logloss: 0.651968 feature_fraction_stage2, val_score: 0.649769: 83%|########3 | 5/6 [00:03<00:00, 1.53it/s][I 2020-09-27 04:43:41,380] Trial 41 finished with value: 0.6519681937770695 and parameters: {'feature_fraction': 0.9159999999999999}. Best is trial 37 with value: 0.6506622068727096. feature_fraction_stage2, val_score: 0.649769: 83%|########3 | 5/6 [00:03<00:00, 1.53it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005114 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.607358 valid's binary_logloss: 0.650129 [200] train's binary_logloss: 0.572939 valid's binary_logloss: 0.650833 Early stopping, best iteration is: [137] train's binary_logloss: 0.593957 valid's binary_logloss: 0.649769 feature_fraction_stage2, val_score: 0.649769: 100%|##########| 6/6 [00:04<00:00, 1.54it/s][I 2020-09-27 04:43:42,023] Trial 42 finished with value: 0.649768864866138 and parameters: {'feature_fraction': 0.8839999999999999}. Best is trial 42 with value: 0.649768864866138. feature_fraction_stage2, val_score: 0.649769: 100%|##########| 6/6 [00:04<00:00, 1.40it/s] regularization_factors, val_score: 0.649769: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000939 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.607973 valid's binary_logloss: 0.65115 [200] train's binary_logloss: 0.573107 valid's binary_logloss: 0.649872 [300] train's binary_logloss: 0.543329 valid's binary_logloss: 0.651302 Early stopping, best iteration is: [229] train's binary_logloss: 0.564107 valid's binary_logloss: 0.649364 regularization_factors, val_score: 0.649364: 5%|5 | 1/20 [00:01<00:24, 1.27s/it][I 2020-09-27 04:43:43,315] Trial 43 finished with value: 0.6493642189216182 and parameters: {'lambda_l1': 0.16783910794593826, 'lambda_l2': 6.907642804430548e-07}. Best is trial 43 with value: 0.6493642189216182. regularization_factors, val_score: 0.649364: 5%|5 | 1/20 [00:01<00:24, 1.27s/it][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000994 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.608181 valid's binary_logloss: 0.65244 [200] train's binary_logloss: 0.574199 valid's binary_logloss: 0.652483 Early stopping, best iteration is: [119] train's binary_logloss: 0.601088 valid's binary_logloss: 0.651426 regularization_factors, val_score: 0.649364: 10%|# | 2/20 [00:01<00:19, 1.08s/it][I 2020-09-27 04:43:43,949] Trial 44 finished with value: 0.6514259112904369 and parameters: {'lambda_l1': 0.3129099140173397, 'lambda_l2': 3.3514640861329857e-07}. Best is trial 43 with value: 0.6493642189216182. regularization_factors, val_score: 0.649364: 10%|# | 2/20 [00:01<00:19, 1.08s/it][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009984 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.607688 valid's binary_logloss: 0.651501 [200] train's binary_logloss: 0.57304 valid's binary_logloss: 0.648731 [300] train's binary_logloss: 0.543627 valid's binary_logloss: 0.650347 Early stopping, best iteration is: [212] train's binary_logloss: 0.569524 valid's binary_logloss: 0.64867 regularization_factors, val_score: 0.648670: 15%|#5 | 3/20 [00:02<00:16, 1.01it/s][I 2020-09-27 04:43:44,720] Trial 45 finished with value: 0.6486701923102753 and parameters: {'lambda_l1': 7.774361825486131e-06, 'lambda_l2': 0.006013598728332988}. Best is trial 45 with value: 0.6486701923102753. regularization_factors, val_score: 0.648670: 15%|#5 | 3/20 [00:02<00:16, 1.01it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000855 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.607956 valid's binary_logloss: 0.650205 [200] train's binary_logloss: 0.573144 valid's binary_logloss: 0.651349 Early stopping, best iteration is: [109] train's binary_logloss: 0.604467 valid's binary_logloss: 0.649709 regularization_factors, val_score: 0.648670: 20%|## | 4/20 [00:03<00:13, 1.17it/s][I 2020-09-27 04:43:45,259] Trial 46 finished with value: 0.6497086221332595 and parameters: {'lambda_l1': 4.0595395206550105e-07, 'lambda_l2': 0.16859061913783596}. Best is trial 45 with value: 0.6486701923102753. regularization_factors, val_score: 0.648670: 20%|## | 4/20 [00:03<00:13, 1.17it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000898 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.607896 valid's binary_logloss: 0.649423 [200] train's binary_logloss: 0.574175 valid's binary_logloss: 0.64969 Early stopping, best iteration is: [121] train's binary_logloss: 0.600172 valid's binary_logloss: 0.648913 regularization_factors, val_score: 0.648670: 25%|##5 | 5/20 [00:03<00:11, 1.29it/s][I 2020-09-27 04:43:45,841] Trial 47 finished with value: 0.6489130270973635 and parameters: {'lambda_l1': 1.9089796516560767e-07, 'lambda_l2': 0.2199091551428564}. Best is trial 45 with value: 0.6486701923102753. regularization_factors, val_score: 0.648670: 25%|##5 | 5/20 [00:03<00:11, 1.29it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000906 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.607889 valid's binary_logloss: 0.650079 [200] train's binary_logloss: 0.574212 valid's binary_logloss: 0.65031 Early stopping, best iteration is: [131] train's binary_logloss: 0.59635 valid's binary_logloss: 0.648762 regularization_factors, val_score: 0.648670: 30%|### | 6/20 [00:04<00:11, 1.23it/s][I 2020-09-27 04:43:46,751] Trial 48 finished with value: 0.6487624267173515 and parameters: {'lambda_l1': 4.2319368851060655e-07, 'lambda_l2': 0.34004310864421733}. Best is trial 45 with value: 0.6486701923102753. regularization_factors, val_score: 0.648670: 30%|### | 6/20 [00:04<00:11, 1.23it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.014681 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.608275 valid's binary_logloss: 0.650944 [200] train's binary_logloss: 0.573416 valid's binary_logloss: 0.651579 Early stopping, best iteration is: [113] train's binary_logloss: 0.603189 valid's binary_logloss: 0.650489 regularization_factors, val_score: 0.648670: 35%|###5 | 7/20 [00:05<00:10, 1.22it/s][I 2020-09-27 04:43:47,591] Trial 49 finished with value: 0.6504888821768885 and parameters: {'lambda_l1': 1.875716331797824e-07, 'lambda_l2': 0.3304208617133868}. Best is trial 45 with value: 0.6486701923102753. regularization_factors, val_score: 0.648670: 35%|###5 | 7/20 [00:05<00:10, 1.22it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001188 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.608185 valid's binary_logloss: 0.650237 Early stopping, best iteration is: [98] train's binary_logloss: 0.608923 valid's binary_logloss: 0.650108 regularization_factors, val_score: 0.648670: 40%|#### | 8/20 [00:06<00:08, 1.35it/s][I 2020-09-27 04:43:48,143] Trial 50 finished with value: 0.6501084626471252 and parameters: {'lambda_l1': 4.671913690463379e-07, 'lambda_l2': 0.1995460595674108}. Best is trial 45 with value: 0.6486701923102753. regularization_factors, val_score: 0.648670: 40%|#### | 8/20 [00:06<00:08, 1.35it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000823 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.60758 valid's binary_logloss: 0.651374 [200] train's binary_logloss: 0.57306 valid's binary_logloss: 0.651712 Early stopping, best iteration is: [154] train's binary_logloss: 0.588205 valid's binary_logloss: 0.650518 regularization_factors, val_score: 0.648670: 45%|####5 | 9/20 [00:06<00:07, 1.39it/s][I 2020-09-27 04:43:48,805] Trial 51 finished with value: 0.6505179255791086 and parameters: {'lambda_l1': 7.519476866197216e-07, 'lambda_l2': 0.08987068172093761}. Best is trial 45 with value: 0.6486701923102753. regularization_factors, val_score: 0.648670: 45%|####5 | 9/20 [00:06<00:07, 1.39it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000917 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.607423 valid's binary_logloss: 0.652085 [200] train's binary_logloss: 0.573345 valid's binary_logloss: 0.651502 Early stopping, best iteration is: [183] train's binary_logloss: 0.578885 valid's binary_logloss: 0.651003 regularization_factors, val_score: 0.648670: 50%|##### | 10/20 [00:07<00:07, 1.40it/s][I 2020-09-27 04:43:49,511] Trial 52 finished with value: 0.6510031934822363 and parameters: {'lambda_l1': 7.307330460252661e-07, 'lambda_l2': 0.0186645031244208}. Best is trial 45 with value: 0.6486701923102753. regularization_factors, val_score: 0.648670: 50%|##### | 10/20 [00:07<00:07, 1.40it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002545 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.607358 valid's binary_logloss: 0.650129 [200] train's binary_logloss: 0.572941 valid's binary_logloss: 0.650864 Early stopping, best iteration is: [137] train's binary_logloss: 0.593957 valid's binary_logloss: 0.649769 regularization_factors, val_score: 0.648670: 55%|#####5 | 11/20 [00:08<00:06, 1.46it/s][I 2020-09-27 04:43:50,135] Trial 53 finished with value: 0.6497687989983096 and parameters: {'lambda_l1': 0.0002292384736520708, 'lambda_l2': 8.689375650085261e-05}. Best is trial 45 with value: 0.6486701923102753. regularization_factors, val_score: 0.648670: 55%|#####5 | 11/20 [00:08<00:06, 1.46it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000469 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.607679 valid's binary_logloss: 0.651246 [200] train's binary_logloss: 0.57339 valid's binary_logloss: 0.653045 Early stopping, best iteration is: [127] train's binary_logloss: 0.597662 valid's binary_logloss: 0.650697 regularization_factors, val_score: 0.648670: 60%|###### | 12/20 [00:09<00:06, 1.23it/s][I 2020-09-27 04:43:51,237] Trial 54 finished with value: 0.6506972734005597 and parameters: {'lambda_l1': 0.001430301077955183, 'lambda_l2': 5.329476633929677e-06}. Best is trial 45 with value: 0.6486701923102753. regularization_factors, val_score: 0.648670: 60%|###### | 12/20 [00:09<00:06, 1.23it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000893 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.607359 valid's binary_logloss: 0.650129 [200] train's binary_logloss: 0.572942 valid's binary_logloss: 0.650833 Early stopping, best iteration is: [137] train's binary_logloss: 0.593959 valid's binary_logloss: 0.649769 regularization_factors, val_score: 0.648670: 65%|######5 | 13/20 [00:09<00:05, 1.31it/s][I 2020-09-27 04:43:51,881] Trial 55 finished with value: 0.6497686874558546 and parameters: {'lambda_l1': 0.00011580371818856396, 'lambda_l2': 0.0005902239811498276}. Best is trial 45 with value: 0.6486701923102753. regularization_factors, val_score: 0.648670: 65%|######5 | 13/20 [00:09<00:05, 1.31it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.007667 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.607682 valid's binary_logloss: 0.651246 [200] train's binary_logloss: 0.57384 valid's binary_logloss: 0.650954 [300] train's binary_logloss: 0.543099 valid's binary_logloss: 0.650065 Early stopping, best iteration is: [273] train's binary_logloss: 0.551078 valid's binary_logloss: 0.649674 regularization_factors, val_score: 0.648670: 70%|####### | 14/20 [00:10<00:04, 1.25it/s][I 2020-09-27 04:43:52,762] Trial 56 finished with value: 0.649674471194536 and parameters: {'lambda_l1': 3.268557145355658e-05, 'lambda_l2': 0.003069238534147287}. Best is trial 45 with value: 0.6486701923102753. regularization_factors, val_score: 0.648670: 70%|####### | 14/20 [00:10<00:04, 1.25it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000897 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.612768 valid's binary_logloss: 0.651575 [200] train's binary_logloss: 0.586233 valid's binary_logloss: 0.651305 Early stopping, best iteration is: [160] train's binary_logloss: 0.595921 valid's binary_logloss: 0.650515 regularization_factors, val_score: 0.648670: 75%|#######5 | 15/20 [00:11<00:03, 1.31it/s][I 2020-09-27 04:43:53,439] Trial 57 finished with value: 0.6505145838820116 and parameters: {'lambda_l1': 5.615730195745978e-06, 'lambda_l2': 7.365468833430753}. Best is trial 45 with value: 0.6486701923102753. regularization_factors, val_score: 0.648670: 75%|#######5 | 15/20 [00:11<00:03, 1.31it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.007481 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.60768 valid's binary_logloss: 0.651246 [200] train's binary_logloss: 0.573394 valid's binary_logloss: 0.653045 Early stopping, best iteration is: [127] train's binary_logloss: 0.597664 valid's binary_logloss: 0.650697 regularization_factors, val_score: 0.648670: 80%|######## | 16/20 [00:12<00:02, 1.38it/s][I 2020-09-27 04:43:54,076] Trial 58 finished with value: 0.6506971138019576 and parameters: {'lambda_l1': 2.3099984324631693e-08, 'lambda_l2': 0.0023773905893283293}. Best is trial 45 with value: 0.6486701923102753. regularization_factors, val_score: 0.648670: 80%|######## | 16/20 [00:12<00:02, 1.38it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001747 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.61046 valid's binary_logloss: 0.65107 [200] train's binary_logloss: 0.581203 valid's binary_logloss: 0.651453 Early stopping, best iteration is: [171] train's binary_logloss: 0.588919 valid's binary_logloss: 0.650336 regularization_factors, val_score: 0.648670: 85%|########5 | 17/20 [00:13<00:02, 1.15it/s][I 2020-09-27 04:43:55,278] Trial 59 finished with value: 0.6503360249834589 and parameters: {'lambda_l1': 1.2365048737402212e-05, 'lambda_l2': 3.1664647322579405}. Best is trial 45 with value: 0.6486701923102753. regularization_factors, val_score: 0.648670: 85%|########5 | 17/20 [00:13<00:02, 1.15it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000874 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.607694 valid's binary_logloss: 0.6515 [200] train's binary_logloss: 0.573475 valid's binary_logloss: 0.651154 Early stopping, best iteration is: [133] train's binary_logloss: 0.595863 valid's binary_logloss: 0.650792 regularization_factors, val_score: 0.648670: 90%|######### | 18/20 [00:13<00:01, 1.27it/s][I 2020-09-27 04:43:55,883] Trial 60 finished with value: 0.6507917448366202 and parameters: {'lambda_l1': 1.4432590771728913e-08, 'lambda_l2': 0.01029217837602668}. Best is trial 45 with value: 0.6486701923102753. regularization_factors, val_score: 0.648670: 90%|######### | 18/20 [00:13<00:01, 1.27it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000902 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.607359 valid's binary_logloss: 0.650129 [200] train's binary_logloss: 0.572942 valid's binary_logloss: 0.650864 Early stopping, best iteration is: [137] train's binary_logloss: 0.593958 valid's binary_logloss: 0.649769 regularization_factors, val_score: 0.648670: 95%|#########5| 19/20 [00:14<00:00, 1.34it/s][I 2020-09-27 04:43:56,523] Trial 61 finished with value: 0.6497687138706673 and parameters: {'lambda_l1': 0.00010150982895717758, 'lambda_l2': 0.0005002352741242428}. Best is trial 45 with value: 0.6486701923102753. regularization_factors, val_score: 0.648670: 95%|#########5| 19/20 [00:14<00:00, 1.34it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000964 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.607358 valid's binary_logloss: 0.650129 [200] train's binary_logloss: 0.57294 valid's binary_logloss: 0.650849 Early stopping, best iteration is: [137] train's binary_logloss: 0.593957 valid's binary_logloss: 0.649769 regularization_factors, val_score: 0.648670: 100%|##########| 20/20 [00:15<00:00, 1.42it/s][I 2020-09-27 04:43:57,130] Trial 62 finished with value: 0.6497688040658035 and parameters: {'lambda_l1': 1.4113098720716957e-05, 'lambda_l2': 0.00022029341576689507}. Best is trial 45 with value: 0.6486701923102753. regularization_factors, val_score: 0.648670: 100%|##########| 20/20 [00:15<00:00, 1.32it/s] min_data_in_leaf, val_score: 0.648670: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000908 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.607864 valid's binary_logloss: 0.650842 [200] train's binary_logloss: 0.573843 valid's binary_logloss: 0.649839 [300] train's binary_logloss: 0.544181 valid's binary_logloss: 0.65183 Early stopping, best iteration is: [206] train's binary_logloss: 0.572189 valid's binary_logloss: 0.649438 min_data_in_leaf, val_score: 0.648670: 20%|## | 1/5 [00:00<00:03, 1.26it/s][I 2020-09-27 04:43:57,935] Trial 63 finished with value: 0.6494379643673076 and parameters: {'min_child_samples': 25}. Best is trial 63 with value: 0.6494379643673076. min_data_in_leaf, val_score: 0.648670: 20%|## | 1/5 [00:00<00:03, 1.26it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012759 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.610343 valid's binary_logloss: 0.652543 Early stopping, best iteration is: [69] train's binary_logloss: 0.622017 valid's binary_logloss: 0.652141 min_data_in_leaf, val_score: 0.648670: 40%|#### | 2/5 [00:01<00:02, 1.13it/s][I 2020-09-27 04:43:59,052] Trial 64 finished with value: 0.6521414167861022 and parameters: {'min_child_samples': 100}. Best is trial 63 with value: 0.6494379643673076. min_data_in_leaf, val_score: 0.648670: 40%|#### | 2/5 [00:01<00:02, 1.13it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000904 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.60911 valid's binary_logloss: 0.649506 [200] train's binary_logloss: 0.576093 valid's binary_logloss: 0.651061 Early stopping, best iteration is: [123] train's binary_logloss: 0.601092 valid's binary_logloss: 0.648901 min_data_in_leaf, val_score: 0.648670: 60%|###### | 3/5 [00:02<00:01, 1.23it/s][I 2020-09-27 04:43:59,681] Trial 65 finished with value: 0.6489009066838253 and parameters: {'min_child_samples': 50}. Best is trial 65 with value: 0.6489009066838253. min_data_in_leaf, val_score: 0.648670: 60%|###### | 3/5 [00:02<00:01, 1.23it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000900 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.607172 valid's binary_logloss: 0.65 [200] train's binary_logloss: 0.572515 valid's binary_logloss: 0.651594 Early stopping, best iteration is: [100] train's binary_logloss: 0.607172 valid's binary_logloss: 0.65 min_data_in_leaf, val_score: 0.648670: 80%|######## | 4/5 [00:03<00:00, 1.33it/s][I 2020-09-27 04:44:00,293] Trial 66 finished with value: 0.6500004234119392 and parameters: {'min_child_samples': 10}. Best is trial 65 with value: 0.6489009066838253. min_data_in_leaf, val_score: 0.648670: 80%|######## | 4/5 [00:03<00:00, 1.33it/s][LightGBM] [Info] Number of positive: 12813, number of negative: 13186 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000902 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4239 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.492827 -> initscore=-0.028695 [LightGBM] [Info] Start training from score -0.028695 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.607283 valid's binary_logloss: 0.652029 [200] train's binary_logloss: 0.572898 valid's binary_logloss: 0.653554 Early stopping, best iteration is: [137] train's binary_logloss: 0.593598 valid's binary_logloss: 0.650833 min_data_in_leaf, val_score: 0.648670: 100%|##########| 5/5 [00:03<00:00, 1.38it/s][I 2020-09-27 04:44:00,964] Trial 67 finished with value: 0.6508328689277415 and parameters: {'min_child_samples': 5}. Best is trial 65 with value: 0.6489009066838253. min_data_in_leaf, val_score: 0.648670: 100%|##########| 5/5 [00:03<00:00, 1.31it/s] Fold : 6 [I 2020-09-27 04:44:01,016] A new study created in memory with name: no-name-d7eb95f7-845c-40a3-9583-af27505411ce feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.003088 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.574382 valid's binary_logloss: 0.658832 Early stopping, best iteration is: [54] train's binary_logloss: 0.607296 valid's binary_logloss: 0.657769 feature_fraction, val_score: 0.657769: 14%|#4 | 1/7 [00:00<00:03, 1.69it/s][I 2020-09-27 04:44:01,616] Trial 0 finished with value: 0.6577689712084953 and parameters: {'feature_fraction': 0.7}. Best is trial 0 with value: 0.6577689712084953. feature_fraction, val_score: 0.657769: 14%|#4 | 1/7 [00:00<00:03, 1.69it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004999 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.573426 valid's binary_logloss: 0.659195 Early stopping, best iteration is: [52] train's binary_logloss: 0.607667 valid's binary_logloss: 0.65734 feature_fraction, val_score: 0.657340: 29%|##8 | 2/7 [00:01<00:03, 1.50it/s][I 2020-09-27 04:44:02,465] Trial 1 finished with value: 0.6573399663499826 and parameters: {'feature_fraction': 0.8}. Best is trial 1 with value: 0.6573399663499826. feature_fraction, val_score: 0.657340: 29%|##8 | 2/7 [00:01<00:03, 1.50it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012514 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.578959 valid's binary_logloss: 0.658522 Early stopping, best iteration is: [92] train's binary_logloss: 0.583977 valid's binary_logloss: 0.657645 feature_fraction, val_score: 0.657340: 43%|####2 | 3/7 [00:02<00:02, 1.48it/s][I 2020-09-27 04:44:03,161] Trial 2 finished with value: 0.6576453641626878 and parameters: {'feature_fraction': 0.5}. Best is trial 1 with value: 0.6573399663499826. feature_fraction, val_score: 0.657340: 43%|####2 | 3/7 [00:02<00:02, 1.48it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000380 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.582749 valid's binary_logloss: 0.655485 [200] train's binary_logloss: 0.531378 valid's binary_logloss: 0.660147 Early stopping, best iteration is: [107] train's binary_logloss: 0.57849 valid's binary_logloss: 0.654569 feature_fraction, val_score: 0.654569: 57%|#####7 | 4/7 [00:03<00:02, 1.04it/s][I 2020-09-27 04:44:04,801] Trial 3 finished with value: 0.654569007042944 and parameters: {'feature_fraction': 0.4}. Best is trial 3 with value: 0.654569007042944. feature_fraction, val_score: 0.654569: 57%|#####7 | 4/7 [00:03<00:02, 1.04it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000972 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.572042 valid's binary_logloss: 0.658846 Early stopping, best iteration is: [79] train's binary_logloss: 0.586097 valid's binary_logloss: 0.65804 feature_fraction, val_score: 0.654569: 71%|#######1 | 5/7 [00:04<00:01, 1.16it/s][I 2020-09-27 04:44:05,420] Trial 4 finished with value: 0.6580401508293939 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 3 with value: 0.654569007042944. feature_fraction, val_score: 0.654569: 71%|#######1 | 5/7 [00:04<00:01, 1.16it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000929 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.571857 valid's binary_logloss: 0.659029 Early stopping, best iteration is: [44] train's binary_logloss: 0.612853 valid's binary_logloss: 0.657339 feature_fraction, val_score: 0.654569: 86%|########5 | 6/7 [00:04<00:00, 1.30it/s][I 2020-09-27 04:44:05,979] Trial 5 finished with value: 0.6573389310156069 and parameters: {'feature_fraction': 1.0}. Best is trial 3 with value: 0.654569007042944. feature_fraction, val_score: 0.654569: 86%|########5 | 6/7 [00:04<00:00, 1.30it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016710 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.577175 valid's binary_logloss: 0.657844 Early stopping, best iteration is: [54] train's binary_logloss: 0.608418 valid's binary_logloss: 0.656152 feature_fraction, val_score: 0.654569: 100%|##########| 7/7 [00:05<00:00, 1.24it/s][I 2020-09-27 04:44:06,878] Trial 6 finished with value: 0.6561524021651702 and parameters: {'feature_fraction': 0.6}. Best is trial 3 with value: 0.654569007042944. feature_fraction, val_score: 0.654569: 100%|##########| 7/7 [00:05<00:00, 1.19it/s] num_leaves, val_score: 0.654569: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004746 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.384415 valid's binary_logloss: 0.6703 Early stopping, best iteration is: [44] train's binary_logloss: 0.508564 valid's binary_logloss: 0.662989 num_leaves, val_score: 0.654569: 5%|5 | 1/20 [00:00<00:14, 1.28it/s][I 2020-09-27 04:44:07,670] Trial 7 finished with value: 0.6629889780799463 and parameters: {'num_leaves': 164}. Best is trial 7 with value: 0.6629889780799463. num_leaves, val_score: 0.654569: 5%|5 | 1/20 [00:00<00:14, 1.28it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000332 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.410682 valid's binary_logloss: 0.66176 Early stopping, best iteration is: [62] train's binary_logloss: 0.482134 valid's binary_logloss: 0.658251 num_leaves, val_score: 0.654569: 10%|# | 2/20 [00:01<00:14, 1.22it/s][I 2020-09-27 04:44:08,569] Trial 8 finished with value: 0.6582514287091177 and parameters: {'num_leaves': 142}. Best is trial 8 with value: 0.6582514287091177. num_leaves, val_score: 0.654569: 10%|# | 2/20 [00:01<00:14, 1.22it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000603 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.412047 valid's binary_logloss: 0.659767 Early stopping, best iteration is: [58] train's binary_logloss: 0.492004 valid's binary_logloss: 0.656213 num_leaves, val_score: 0.654569: 15%|#5 | 3/20 [00:02<00:14, 1.20it/s][I 2020-09-27 04:44:09,432] Trial 9 finished with value: 0.6562133173559287 and parameters: {'num_leaves': 140}. Best is trial 9 with value: 0.6562133173559287. num_leaves, val_score: 0.654569: 15%|#5 | 3/20 [00:02<00:14, 1.20it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004277 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.608851 valid's binary_logloss: 0.656989 Early stopping, best iteration is: [87] train's binary_logloss: 0.614406 valid's binary_logloss: 0.656446 num_leaves, val_score: 0.654569: 20%|## | 4/20 [00:03<00:11, 1.37it/s][I 2020-09-27 04:44:09,931] Trial 10 finished with value: 0.6564457093023696 and parameters: {'num_leaves': 19}. Best is trial 9 with value: 0.6562133173559287. num_leaves, val_score: 0.654569: 20%|## | 4/20 [00:03<00:11, 1.37it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004598 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.301686 valid's binary_logloss: 0.673949 Early stopping, best iteration is: [31] train's binary_logloss: 0.504321 valid's binary_logloss: 0.666042 num_leaves, val_score: 0.654569: 25%|##5 | 5/20 [00:04<00:13, 1.09it/s][I 2020-09-27 04:44:11,277] Trial 11 finished with value: 0.66604185972251 and parameters: {'num_leaves': 254}. Best is trial 9 with value: 0.6562133173559287. num_leaves, val_score: 0.654569: 25%|##5 | 5/20 [00:04<00:13, 1.09it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000377 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.618907 valid's binary_logloss: 0.654583 [200] train's binary_logloss: 0.590148 valid's binary_logloss: 0.653268 Early stopping, best iteration is: [171] train's binary_logloss: 0.597901 valid's binary_logloss: 0.652352 num_leaves, val_score: 0.652352: 30%|### | 6/20 [00:04<00:11, 1.24it/s][I 2020-09-27 04:44:11,818] Trial 12 finished with value: 0.6523523176355726 and parameters: {'num_leaves': 15}. Best is trial 12 with value: 0.6523523176355726. num_leaves, val_score: 0.652352: 30%|### | 6/20 [00:04<00:11, 1.24it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000404 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.658358 valid's binary_logloss: 0.663406 [200] train's binary_logloss: 0.647972 valid's binary_logloss: 0.655898 [300] train's binary_logloss: 0.642797 valid's binary_logloss: 0.652806 [400] train's binary_logloss: 0.639134 valid's binary_logloss: 0.651593 [500] train's binary_logloss: 0.636175 valid's binary_logloss: 0.651137 [600] train's binary_logloss: 0.633356 valid's binary_logloss: 0.65085 [700] train's binary_logloss: 0.630829 valid's binary_logloss: 0.650912 Early stopping, best iteration is: [631] train's binary_logloss: 0.63254 valid's binary_logloss: 0.650592 num_leaves, val_score: 0.650592: 35%|###5 | 7/20 [00:05<00:11, 1.17it/s][I 2020-09-27 04:44:12,795] Trial 13 finished with value: 0.6505921558008332 and parameters: {'num_leaves': 3}. Best is trial 13 with value: 0.6505921558008332. num_leaves, val_score: 0.650592: 35%|###5 | 7/20 [00:05<00:11, 1.17it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004382 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.643838 valid's binary_logloss: 0.655968 [200] train's binary_logloss: 0.630606 valid's binary_logloss: 0.651656 [300] train's binary_logloss: 0.62092 valid's binary_logloss: 0.651649 Early stopping, best iteration is: [275] train's binary_logloss: 0.623153 valid's binary_logloss: 0.651151 num_leaves, val_score: 0.650592: 40%|#### | 8/20 [00:06<00:09, 1.29it/s][I 2020-09-27 04:44:13,393] Trial 14 finished with value: 0.6511512029998484 and parameters: {'num_leaves': 6}. Best is trial 13 with value: 0.6505921558008332. num_leaves, val_score: 0.650592: 40%|#### | 8/20 [00:06<00:09, 1.29it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005223 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.536652 valid's binary_logloss: 0.654767 Early stopping, best iteration is: [88] train's binary_logloss: 0.548536 valid's binary_logloss: 0.654093 num_leaves, val_score: 0.650592: 45%|####5 | 9/20 [00:07<00:09, 1.12it/s][I 2020-09-27 04:44:14,546] Trial 15 finished with value: 0.6540930395271628 and parameters: {'num_leaves': 55}. Best is trial 13 with value: 0.6505921558008332. num_leaves, val_score: 0.650592: 45%|####5 | 9/20 [00:07<00:09, 1.12it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000391 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.522391 valid's binary_logloss: 0.660187 Early stopping, best iteration is: [56] train's binary_logloss: 0.573263 valid's binary_logloss: 0.659042 num_leaves, val_score: 0.650592: 50%|##### | 10/20 [00:08<00:07, 1.25it/s][I 2020-09-27 04:44:15,132] Trial 16 finished with value: 0.6590416618563378 and parameters: {'num_leaves': 63}. Best is trial 13 with value: 0.6505921558008332. num_leaves, val_score: 0.650592: 50%|##### | 10/20 [00:08<00:07, 1.25it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000241 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.500962 valid's binary_logloss: 0.659121 Early stopping, best iteration is: [60] train's binary_logloss: 0.552156 valid's binary_logloss: 0.658112 num_leaves, val_score: 0.650592: 55%|#####5 | 11/20 [00:08<00:06, 1.33it/s][I 2020-09-27 04:44:15,768] Trial 17 finished with value: 0.6581124834744749 and parameters: {'num_leaves': 77}. Best is trial 13 with value: 0.6505921558008332. num_leaves, val_score: 0.650592: 55%|#####5 | 11/20 [00:08<00:06, 1.33it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000455 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.626446 valid's binary_logloss: 0.654342 [200] train's binary_logloss: 0.602897 valid's binary_logloss: 0.654305 Early stopping, best iteration is: [128] train's binary_logloss: 0.619208 valid's binary_logloss: 0.653208 num_leaves, val_score: 0.650592: 60%|###### | 12/20 [00:09<00:05, 1.50it/s][I 2020-09-27 04:44:16,240] Trial 18 finished with value: 0.6532083579968055 and parameters: {'num_leaves': 12}. Best is trial 13 with value: 0.6505921558008332. num_leaves, val_score: 0.650592: 60%|###### | 12/20 [00:09<00:05, 1.50it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000331 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.340384 valid's binary_logloss: 0.669675 Early stopping, best iteration is: [43] train's binary_logloss: 0.483746 valid's binary_logloss: 0.660937 num_leaves, val_score: 0.650592: 65%|######5 | 13/20 [00:10<00:05, 1.31it/s][I 2020-09-27 04:44:17,237] Trial 19 finished with value: 0.6609367707387375 and parameters: {'num_leaves': 209}. Best is trial 13 with value: 0.6505921558008332. num_leaves, val_score: 0.650592: 65%|######5 | 13/20 [00:10<00:05, 1.31it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000416 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.47136 valid's binary_logloss: 0.661857 Early stopping, best iteration is: [66] train's binary_logloss: 0.520512 valid's binary_logloss: 0.659368 num_leaves, val_score: 0.650592: 70%|####### | 14/20 [00:11<00:05, 1.08it/s][I 2020-09-27 04:44:18,525] Trial 20 finished with value: 0.6593681790082808 and parameters: {'num_leaves': 95}. Best is trial 13 with value: 0.6505921558008332. num_leaves, val_score: 0.650592: 70%|####### | 14/20 [00:11<00:05, 1.08it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000434 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.634417 valid's binary_logloss: 0.655116 [200] train's binary_logloss: 0.616494 valid's binary_logloss: 0.653099 [300] train's binary_logloss: 0.601536 valid's binary_logloss: 0.65265 Early stopping, best iteration is: [250] train's binary_logloss: 0.608783 valid's binary_logloss: 0.651395 num_leaves, val_score: 0.650592: 75%|#######5 | 15/20 [00:12<00:04, 1.20it/s][I 2020-09-27 04:44:19,142] Trial 21 finished with value: 0.6513954160672126 and parameters: {'num_leaves': 9}. Best is trial 13 with value: 0.6505921558008332. num_leaves, val_score: 0.650592: 75%|#######5 | 15/20 [00:12<00:04, 1.20it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000366 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.668969 valid's binary_logloss: 0.671303 [200] train's binary_logloss: 0.659328 valid's binary_logloss: 0.662192 [300] train's binary_logloss: 0.65408 valid's binary_logloss: 0.657049 [400] train's binary_logloss: 0.650963 valid's binary_logloss: 0.654645 [500] train's binary_logloss: 0.648956 valid's binary_logloss: 0.652696 [600] train's binary_logloss: 0.647608 valid's binary_logloss: 0.651783 [700] train's binary_logloss: 0.646665 valid's binary_logloss: 0.651143 [800] train's binary_logloss: 0.645958 valid's binary_logloss: 0.650652 [900] train's binary_logloss: 0.645405 valid's binary_logloss: 0.6504 [1000] train's binary_logloss: 0.644946 valid's binary_logloss: 0.650075 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.644946 valid's binary_logloss: 0.650075 num_leaves, val_score: 0.650075: 80%|######## | 16/20 [00:13<00:03, 1.05it/s][I 2020-09-27 04:44:20,393] Trial 22 finished with value: 0.6500754523975805 and parameters: {'num_leaves': 2}. Best is trial 22 with value: 0.6500754523975805. num_leaves, val_score: 0.650075: 80%|######## | 16/20 [00:13<00:03, 1.05it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.005069 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.565299 valid's binary_logloss: 0.658702 Early stopping, best iteration is: [60] train's binary_logloss: 0.597675 valid's binary_logloss: 0.657428 num_leaves, val_score: 0.650075: 85%|########5 | 17/20 [00:13<00:02, 1.23it/s][I 2020-09-27 04:44:20,871] Trial 23 finished with value: 0.6574278070760912 and parameters: {'num_leaves': 39}. Best is trial 22 with value: 0.6500754523975805. num_leaves, val_score: 0.650075: 85%|########5 | 17/20 [00:13<00:02, 1.23it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000464 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.652156 valid's binary_logloss: 0.66018 [200] train's binary_logloss: 0.641398 valid's binary_logloss: 0.653863 [300] train's binary_logloss: 0.635014 valid's binary_logloss: 0.652565 [400] train's binary_logloss: 0.629702 valid's binary_logloss: 0.65154 [500] train's binary_logloss: 0.624786 valid's binary_logloss: 0.652002 Early stopping, best iteration is: [409] train's binary_logloss: 0.629299 valid's binary_logloss: 0.651463 num_leaves, val_score: 0.650075: 90%|######### | 18/20 [00:15<00:01, 1.05it/s][I 2020-09-27 04:44:22,155] Trial 24 finished with value: 0.6514630998860335 and parameters: {'num_leaves': 4}. Best is trial 22 with value: 0.6500754523975805. num_leaves, val_score: 0.650075: 90%|######### | 18/20 [00:15<00:01, 1.05it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000333 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.570374 valid's binary_logloss: 0.656731 Early stopping, best iteration is: [72] train's binary_logloss: 0.590744 valid's binary_logloss: 0.655417 num_leaves, val_score: 0.650075: 95%|#########5| 19/20 [00:15<00:00, 1.23it/s][I 2020-09-27 04:44:22,639] Trial 25 finished with value: 0.655416641463825 and parameters: {'num_leaves': 37}. Best is trial 22 with value: 0.6500754523975805. num_leaves, val_score: 0.650075: 95%|#########5| 19/20 [00:15<00:00, 1.23it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000693 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.455281 valid's binary_logloss: 0.66208 Early stopping, best iteration is: [62] train's binary_logloss: 0.51632 valid's binary_logloss: 0.658037 num_leaves, val_score: 0.650075: 100%|##########| 20/20 [00:16<00:00, 1.26it/s][I 2020-09-27 04:44:23,396] Trial 26 finished with value: 0.6580367008399806 and parameters: {'num_leaves': 107}. Best is trial 22 with value: 0.6500754523975805. num_leaves, val_score: 0.650075: 100%|##########| 20/20 [00:16<00:00, 1.21it/s] bagging, val_score: 0.650075: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000433 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667722 valid's binary_logloss: 0.669888 [200] train's binary_logloss: 0.657469 valid's binary_logloss: 0.660998 [300] train's binary_logloss: 0.652189 valid's binary_logloss: 0.656002 [400] train's binary_logloss: 0.649008 valid's binary_logloss: 0.65329 [500] train's binary_logloss: 0.64724 valid's binary_logloss: 0.651649 [600] train's binary_logloss: 0.646196 valid's binary_logloss: 0.651202 [700] train's binary_logloss: 0.645282 valid's binary_logloss: 0.650767 [800] train's binary_logloss: 0.644572 valid's binary_logloss: 0.650399 [900] train's binary_logloss: 0.643859 valid's binary_logloss: 0.650238 [1000] train's binary_logloss: 0.643295 valid's binary_logloss: 0.649949 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.643295 valid's binary_logloss: 0.649949 bagging, val_score: 0.649949: 10%|# | 1/10 [00:01<00:12, 1.38s/it][I 2020-09-27 04:44:24,791] Trial 27 finished with value: 0.6499492453087289 and parameters: {'bagging_fraction': 0.6374753391797192, 'bagging_freq': 6}. Best is trial 27 with value: 0.6499492453087289. bagging, val_score: 0.649949: 10%|# | 1/10 [00:01<00:12, 1.38s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000379 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667741 valid's binary_logloss: 0.670126 [200] train's binary_logloss: 0.657401 valid's binary_logloss: 0.660759 [300] train's binary_logloss: 0.652134 valid's binary_logloss: 0.655834 [400] train's binary_logloss: 0.649009 valid's binary_logloss: 0.653296 [500] train's binary_logloss: 0.647228 valid's binary_logloss: 0.651516 [600] train's binary_logloss: 0.646169 valid's binary_logloss: 0.65095 [700] train's binary_logloss: 0.645307 valid's binary_logloss: 0.650552 [800] train's binary_logloss: 0.644552 valid's binary_logloss: 0.650135 [900] train's binary_logloss: 0.643876 valid's binary_logloss: 0.650283 Early stopping, best iteration is: [811] train's binary_logloss: 0.64445 valid's binary_logloss: 0.649941 bagging, val_score: 0.649941: 20%|## | 2/10 [00:03<00:11, 1.48s/it][I 2020-09-27 04:44:26,503] Trial 28 finished with value: 0.6499412034252128 and parameters: {'bagging_fraction': 0.6431039528688518, 'bagging_freq': 6}. Best is trial 28 with value: 0.6499412034252128. bagging, val_score: 0.649941: 20%|## | 2/10 [00:03<00:11, 1.48s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000402 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844 [200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602 [300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818 [400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001 [500] train's binary_logloss: 0.647221 valid's binary_logloss: 0.651249 [600] train's binary_logloss: 0.646154 valid's binary_logloss: 0.650938 [700] train's binary_logloss: 0.645276 valid's binary_logloss: 0.650398 [800] train's binary_logloss: 0.644511 valid's binary_logloss: 0.65025 [900] train's binary_logloss: 0.643837 valid's binary_logloss: 0.650107 [1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721 bagging, val_score: 0.649721: 30%|### | 3/10 [00:04<00:10, 1.44s/it][I 2020-09-27 04:44:27,842] Trial 29 finished with value: 0.6497207980729593 and parameters: {'bagging_fraction': 0.6147135266937416, 'bagging_freq': 6}. Best is trial 29 with value: 0.6497207980729593. bagging, val_score: 0.649721: 30%|### | 3/10 [00:04<00:10, 1.44s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000403 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.66762 valid's binary_logloss: 0.670222 [200] train's binary_logloss: 0.657378 valid's binary_logloss: 0.660905 [300] train's binary_logloss: 0.652057 valid's binary_logloss: 0.655995 [400] train's binary_logloss: 0.648925 valid's binary_logloss: 0.653365 [500] train's binary_logloss: 0.647153 valid's binary_logloss: 0.65154 [600] train's binary_logloss: 0.646118 valid's binary_logloss: 0.651221 [700] train's binary_logloss: 0.64522 valid's binary_logloss: 0.650848 [800] train's binary_logloss: 0.644499 valid's binary_logloss: 0.650374 [900] train's binary_logloss: 0.643818 valid's binary_logloss: 0.650522 Early stopping, best iteration is: [811] train's binary_logloss: 0.644399 valid's binary_logloss: 0.650183 bagging, val_score: 0.649721: 40%|#### | 4/10 [00:05<00:08, 1.40s/it][I 2020-09-27 04:44:29,154] Trial 30 finished with value: 0.6501826485329506 and parameters: {'bagging_fraction': 0.6282514585296812, 'bagging_freq': 6}. Best is trial 29 with value: 0.6497207980729593. bagging, val_score: 0.649721: 40%|#### | 4/10 [00:05<00:08, 1.40s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000425 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667636 valid's binary_logloss: 0.670402 [200] train's binary_logloss: 0.65741 valid's binary_logloss: 0.660721 [300] train's binary_logloss: 0.652135 valid's binary_logloss: 0.655888 [400] train's binary_logloss: 0.649005 valid's binary_logloss: 0.653394 [500] train's binary_logloss: 0.647265 valid's binary_logloss: 0.651817 [600] train's binary_logloss: 0.646169 valid's binary_logloss: 0.651429 [700] train's binary_logloss: 0.645289 valid's binary_logloss: 0.651273 Early stopping, best iteration is: [625] train's binary_logloss: 0.645884 valid's binary_logloss: 0.650972 bagging, val_score: 0.649721: 50%|##### | 5/10 [00:07<00:06, 1.39s/it][I 2020-09-27 04:44:30,538] Trial 31 finished with value: 0.6509722145299361 and parameters: {'bagging_fraction': 0.625788067973422, 'bagging_freq': 6}. Best is trial 29 with value: 0.6497207980729593. bagging, val_score: 0.649721: 50%|##### | 5/10 [00:07<00:06, 1.39s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000386 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.66771 valid's binary_logloss: 0.670672 [200] train's binary_logloss: 0.657453 valid's binary_logloss: 0.660978 [300] train's binary_logloss: 0.652108 valid's binary_logloss: 0.655884 [400] train's binary_logloss: 0.648979 valid's binary_logloss: 0.653214 [500] train's binary_logloss: 0.647206 valid's binary_logloss: 0.651492 [600] train's binary_logloss: 0.646131 valid's binary_logloss: 0.650889 [700] train's binary_logloss: 0.645274 valid's binary_logloss: 0.650764 [800] train's binary_logloss: 0.644537 valid's binary_logloss: 0.650318 [900] train's binary_logloss: 0.643867 valid's binary_logloss: 0.650313 Early stopping, best iteration is: [811] train's binary_logloss: 0.644435 valid's binary_logloss: 0.650123 bagging, val_score: 0.649721: 60%|###### | 6/10 [00:08<00:05, 1.34s/it][I 2020-09-27 04:44:31,752] Trial 32 finished with value: 0.650122761267748 and parameters: {'bagging_fraction': 0.6308230415432386, 'bagging_freq': 6}. Best is trial 29 with value: 0.6497207980729593. bagging, val_score: 0.649721: 60%|###### | 6/10 [00:08<00:05, 1.34s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000509 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.668475 valid's binary_logloss: 0.670977 [200] train's binary_logloss: 0.658545 valid's binary_logloss: 0.661641 [300] train's binary_logloss: 0.653145 valid's binary_logloss: 0.656461 [400] train's binary_logloss: 0.649945 valid's binary_logloss: 0.653659 [500] train's binary_logloss: 0.647957 valid's binary_logloss: 0.652033 [600] train's binary_logloss: 0.646663 valid's binary_logloss: 0.651209 [700] train's binary_logloss: 0.645743 valid's binary_logloss: 0.650593 [800] train's binary_logloss: 0.644998 valid's binary_logloss: 0.649891 [900] train's binary_logloss: 0.644366 valid's binary_logloss: 0.649745 [1000] train's binary_logloss: 0.643856 valid's binary_logloss: 0.649763 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.643856 valid's binary_logloss: 0.649763 bagging, val_score: 0.649721: 70%|####### | 7/10 [00:09<00:04, 1.41s/it][I 2020-09-27 04:44:33,330] Trial 33 finished with value: 0.649762914198675 and parameters: {'bagging_fraction': 0.881712034252823, 'bagging_freq': 6}. Best is trial 29 with value: 0.6497207980729593. bagging, val_score: 0.649721: 70%|####### | 7/10 [00:09<00:04, 1.41s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001411 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.668676 valid's binary_logloss: 0.671455 [200] train's binary_logloss: 0.658729 valid's binary_logloss: 0.66187 [300] train's binary_logloss: 0.653368 valid's binary_logloss: 0.656513 [400] train's binary_logloss: 0.650201 valid's binary_logloss: 0.654227 [500] train's binary_logloss: 0.648173 valid's binary_logloss: 0.652343 [600] train's binary_logloss: 0.646868 valid's binary_logloss: 0.651436 [700] train's binary_logloss: 0.645933 valid's binary_logloss: 0.650838 [800] train's binary_logloss: 0.645188 valid's binary_logloss: 0.650137 [900] train's binary_logloss: 0.644607 valid's binary_logloss: 0.64995 [1000] train's binary_logloss: 0.644106 valid's binary_logloss: 0.649864 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.644106 valid's binary_logloss: 0.649864 bagging, val_score: 0.649721: 80%|######## | 8/10 [00:11<00:03, 1.51s/it][I 2020-09-27 04:44:35,075] Trial 34 finished with value: 0.6498641722707447 and parameters: {'bagging_fraction': 0.9222411833757059, 'bagging_freq': 6}. Best is trial 29 with value: 0.6497207980729593. bagging, val_score: 0.649721: 80%|######## | 8/10 [00:11<00:03, 1.51s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000737 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.668902 valid's binary_logloss: 0.671648 [200] train's binary_logloss: 0.65921 valid's binary_logloss: 0.661972 [300] train's binary_logloss: 0.653923 valid's binary_logloss: 0.65701 [400] train's binary_logloss: 0.650779 valid's binary_logloss: 0.654505 [500] train's binary_logloss: 0.648744 valid's binary_logloss: 0.652567 [600] train's binary_logloss: 0.647377 valid's binary_logloss: 0.651548 [700] train's binary_logloss: 0.646422 valid's binary_logloss: 0.651081 [800] train's binary_logloss: 0.645708 valid's binary_logloss: 0.650642 [900] train's binary_logloss: 0.64514 valid's binary_logloss: 0.650392 [1000] train's binary_logloss: 0.644665 valid's binary_logloss: 0.650089 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.644665 valid's binary_logloss: 0.650089 bagging, val_score: 0.649721: 90%|######### | 9/10 [00:13<00:01, 1.50s/it][I 2020-09-27 04:44:36,546] Trial 35 finished with value: 0.6500890749178282 and parameters: {'bagging_fraction': 0.9922180140725776, 'bagging_freq': 2}. Best is trial 29 with value: 0.6497207980729593. bagging, val_score: 0.649721: 90%|######### | 9/10 [00:13<00:01, 1.50s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000379 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.66868 valid's binary_logloss: 0.671198 [200] train's binary_logloss: 0.658732 valid's binary_logloss: 0.661411 [300] train's binary_logloss: 0.653393 valid's binary_logloss: 0.656638 [400] train's binary_logloss: 0.650183 valid's binary_logloss: 0.653991 [500] train's binary_logloss: 0.648188 valid's binary_logloss: 0.652694 [600] train's binary_logloss: 0.646874 valid's binary_logloss: 0.651422 [700] train's binary_logloss: 0.645949 valid's binary_logloss: 0.650978 [800] train's binary_logloss: 0.645235 valid's binary_logloss: 0.650342 [900] train's binary_logloss: 0.644629 valid's binary_logloss: 0.649845 [1000] train's binary_logloss: 0.644118 valid's binary_logloss: 0.649971 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.644118 valid's binary_logloss: 0.649971 bagging, val_score: 0.649721: 100%|##########| 10/10 [00:15<00:00, 1.61s/it][I 2020-09-27 04:44:38,424] Trial 36 finished with value: 0.6499714458134913 and parameters: {'bagging_fraction': 0.9261093121309201, 'bagging_freq': 7}. Best is trial 29 with value: 0.6497207980729593. bagging, val_score: 0.649721: 100%|##########| 10/10 [00:15<00:00, 1.50s/it] feature_fraction_stage2, val_score: 0.649721: 0%| | 0/3 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000513 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667513 valid's binary_logloss: 0.670285 [200] train's binary_logloss: 0.65728 valid's binary_logloss: 0.661217 [300] train's binary_logloss: 0.652006 valid's binary_logloss: 0.656051 [400] train's binary_logloss: 0.648934 valid's binary_logloss: 0.653472 [500] train's binary_logloss: 0.647192 valid's binary_logloss: 0.651779 [600] train's binary_logloss: 0.646099 valid's binary_logloss: 0.651503 [700] train's binary_logloss: 0.645255 valid's binary_logloss: 0.650916 [800] train's binary_logloss: 0.644508 valid's binary_logloss: 0.650651 [900] train's binary_logloss: 0.643868 valid's binary_logloss: 0.650537 Early stopping, best iteration is: [811] train's binary_logloss: 0.644389 valid's binary_logloss: 0.650382 feature_fraction_stage2, val_score: 0.649721: 33%|###3 | 1/3 [00:01<00:02, 1.23s/it][I 2020-09-27 04:44:39,673] Trial 37 finished with value: 0.6503821331647123 and parameters: {'feature_fraction': 0.48000000000000004}. Best is trial 37 with value: 0.6503821331647123. feature_fraction_stage2, val_score: 0.649721: 33%|###3 | 1/3 [00:01<00:02, 1.23s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000502 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667466 valid's binary_logloss: 0.670008 [200] train's binary_logloss: 0.657197 valid's binary_logloss: 0.660612 [300] train's binary_logloss: 0.651972 valid's binary_logloss: 0.655778 [400] train's binary_logloss: 0.648867 valid's binary_logloss: 0.653147 [500] train's binary_logloss: 0.647122 valid's binary_logloss: 0.651573 [600] train's binary_logloss: 0.646035 valid's binary_logloss: 0.650926 [700] train's binary_logloss: 0.645194 valid's binary_logloss: 0.650578 [800] train's binary_logloss: 0.644427 valid's binary_logloss: 0.650523 [900] train's binary_logloss: 0.643736 valid's binary_logloss: 0.650556 Early stopping, best iteration is: [811] train's binary_logloss: 0.644312 valid's binary_logloss: 0.650291 feature_fraction_stage2, val_score: 0.649721: 67%|######6 | 2/3 [00:02<00:01, 1.29s/it][I 2020-09-27 04:44:41,097] Trial 38 finished with value: 0.6502911144681769 and parameters: {'feature_fraction': 0.41600000000000004}. Best is trial 38 with value: 0.6502911144681769. feature_fraction_stage2, val_score: 0.649721: 67%|######6 | 2/3 [00:02<00:01, 1.29s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002477 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667513 valid's binary_logloss: 0.670285 [200] train's binary_logloss: 0.65728 valid's binary_logloss: 0.661217 [300] train's binary_logloss: 0.652006 valid's binary_logloss: 0.656051 [400] train's binary_logloss: 0.648934 valid's binary_logloss: 0.653472 [500] train's binary_logloss: 0.647192 valid's binary_logloss: 0.651779 [600] train's binary_logloss: 0.646099 valid's binary_logloss: 0.651503 [700] train's binary_logloss: 0.645255 valid's binary_logloss: 0.650916 [800] train's binary_logloss: 0.644508 valid's binary_logloss: 0.650651 [900] train's binary_logloss: 0.643868 valid's binary_logloss: 0.650537 Early stopping, best iteration is: [811] train's binary_logloss: 0.644389 valid's binary_logloss: 0.650382 feature_fraction_stage2, val_score: 0.649721: 100%|##########| 3/3 [00:04<00:00, 1.33s/it][I 2020-09-27 04:44:42,520] Trial 39 finished with value: 0.6503821331647123 and parameters: {'feature_fraction': 0.44800000000000006}. Best is trial 38 with value: 0.6502911144681769. feature_fraction_stage2, val_score: 0.649721: 100%|##########| 3/3 [00:04<00:00, 1.36s/it] regularization_factors, val_score: 0.649721: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000379 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844 [200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602 [300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818 [400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001 [500] train's binary_logloss: 0.647221 valid's binary_logloss: 0.651249 [600] train's binary_logloss: 0.646154 valid's binary_logloss: 0.650938 [700] train's binary_logloss: 0.645276 valid's binary_logloss: 0.650398 [800] train's binary_logloss: 0.644511 valid's binary_logloss: 0.65025 [900] train's binary_logloss: 0.643837 valid's binary_logloss: 0.650107 [1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721 regularization_factors, val_score: 0.649721: 5%|5 | 1/20 [00:01<00:25, 1.35s/it][I 2020-09-27 04:44:43,887] Trial 40 finished with value: 0.6497208080532643 and parameters: {'lambda_l1': 4.606489682248236e-06, 'lambda_l2': 0.0004490091062700392}. Best is trial 40 with value: 0.6497208080532643. regularization_factors, val_score: 0.649721: 5%|5 | 1/20 [00:01<00:25, 1.35s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000342 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844 [200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602 [300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818 [400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001 [500] train's binary_logloss: 0.647221 valid's binary_logloss: 0.651249 [600] train's binary_logloss: 0.646154 valid's binary_logloss: 0.650938 [700] train's binary_logloss: 0.645276 valid's binary_logloss: 0.650398 [800] train's binary_logloss: 0.644511 valid's binary_logloss: 0.65025 [900] train's binary_logloss: 0.643837 valid's binary_logloss: 0.650107 [1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721 regularization_factors, val_score: 0.649721: 10%|# | 2/20 [00:03<00:26, 1.50s/it][I 2020-09-27 04:44:45,731] Trial 41 finished with value: 0.6497208177282439 and parameters: {'lambda_l1': 3.367646783477069e-06, 'lambda_l2': 0.0006824013831978261}. Best is trial 40 with value: 0.6497208080532643. regularization_factors, val_score: 0.649721: 10%|# | 2/20 [00:03<00:26, 1.50s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000463 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844 [200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602 [300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818 [400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001 [500] train's binary_logloss: 0.647221 valid's binary_logloss: 0.651249 [600] train's binary_logloss: 0.646154 valid's binary_logloss: 0.650938 [700] train's binary_logloss: 0.645276 valid's binary_logloss: 0.650398 [800] train's binary_logloss: 0.644511 valid's binary_logloss: 0.65025 [900] train's binary_logloss: 0.643837 valid's binary_logloss: 0.650107 [1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721 regularization_factors, val_score: 0.649721: 15%|#5 | 3/20 [00:04<00:24, 1.44s/it][I 2020-09-27 04:44:47,018] Trial 42 finished with value: 0.6497208110181181 and parameters: {'lambda_l1': 1.2594003820953281e-06, 'lambda_l2': 0.0005901706717435754}. Best is trial 40 with value: 0.6497208080532643. regularization_factors, val_score: 0.649721: 15%|#5 | 3/20 [00:04<00:24, 1.44s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000389 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844 [200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602 [300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818 [400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001 [500] train's binary_logloss: 0.647221 valid's binary_logloss: 0.651249 [600] train's binary_logloss: 0.646154 valid's binary_logloss: 0.650938 [700] train's binary_logloss: 0.645276 valid's binary_logloss: 0.650398 [800] train's binary_logloss: 0.644511 valid's binary_logloss: 0.65025 [900] train's binary_logloss: 0.643838 valid's binary_logloss: 0.650107 [1000] train's binary_logloss: 0.643266 valid's binary_logloss: 0.649721 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.643266 valid's binary_logloss: 0.649721 regularization_factors, val_score: 0.649721: 20%|## | 4/20 [00:05<00:22, 1.38s/it][I 2020-09-27 04:44:48,282] Trial 43 finished with value: 0.6497208467241731 and parameters: {'lambda_l1': 1.7272310002902818e-06, 'lambda_l2': 0.0005516895364430477}. Best is trial 40 with value: 0.6497208080532643. regularization_factors, val_score: 0.649721: 20%|## | 4/20 [00:05<00:22, 1.38s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000424 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844 [200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602 [300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818 [400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001 [500] train's binary_logloss: 0.647221 valid's binary_logloss: 0.651249 [600] train's binary_logloss: 0.646154 valid's binary_logloss: 0.650938 [700] train's binary_logloss: 0.645276 valid's binary_logloss: 0.650398 [800] train's binary_logloss: 0.644512 valid's binary_logloss: 0.65025 [900] train's binary_logloss: 0.643838 valid's binary_logloss: 0.650107 [1000] train's binary_logloss: 0.643266 valid's binary_logloss: 0.649721 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.643266 valid's binary_logloss: 0.649721 regularization_factors, val_score: 0.649721: 25%|##5 | 5/20 [00:07<00:22, 1.49s/it][I 2020-09-27 04:44:50,019] Trial 44 finished with value: 0.6497209472919824 and parameters: {'lambda_l1': 1.66122924086294e-06, 'lambda_l2': 0.0004859937346429329}. Best is trial 40 with value: 0.6497208080532643. regularization_factors, val_score: 0.649721: 25%|##5 | 5/20 [00:07<00:22, 1.49s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000580 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844 [200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602 [300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818 [400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001 [500] train's binary_logloss: 0.647221 valid's binary_logloss: 0.651249 [600] train's binary_logloss: 0.646154 valid's binary_logloss: 0.650938 [700] train's binary_logloss: 0.645276 valid's binary_logloss: 0.650398 [800] train's binary_logloss: 0.64451 valid's binary_logloss: 0.65025 [900] train's binary_logloss: 0.643837 valid's binary_logloss: 0.650107 [1000] train's binary_logloss: 0.643264 valid's binary_logloss: 0.649721 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.643264 valid's binary_logloss: 0.649721 regularization_factors, val_score: 0.649721: 30%|### | 6/20 [00:08<00:20, 1.45s/it][I 2020-09-27 04:44:51,378] Trial 45 finished with value: 0.6497207058626523 and parameters: {'lambda_l1': 3.2445067392487893e-06, 'lambda_l2': 0.0004261866691401461}. Best is trial 45 with value: 0.6497207058626523. regularization_factors, val_score: 0.649721: 30%|### | 6/20 [00:08<00:20, 1.45s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000393 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844 [200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602 [300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818 [400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001 [500] train's binary_logloss: 0.647221 valid's binary_logloss: 0.651249 [600] train's binary_logloss: 0.646154 valid's binary_logloss: 0.650938 [700] train's binary_logloss: 0.645276 valid's binary_logloss: 0.650398 [800] train's binary_logloss: 0.644511 valid's binary_logloss: 0.65025 [900] train's binary_logloss: 0.643837 valid's binary_logloss: 0.650107 [1000] train's binary_logloss: 0.643264 valid's binary_logloss: 0.649721 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.643264 valid's binary_logloss: 0.649721 regularization_factors, val_score: 0.649721: 35%|###5 | 7/20 [00:10<00:19, 1.53s/it][I 2020-09-27 04:44:53,078] Trial 46 finished with value: 0.6497207182286581 and parameters: {'lambda_l1': 3.850005827927253e-06, 'lambda_l2': 0.0005152754958015667}. Best is trial 45 with value: 0.6497207058626523. regularization_factors, val_score: 0.649721: 35%|###5 | 7/20 [00:10<00:19, 1.53s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000406 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844 [200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602 [300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818 [400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001 [500] train's binary_logloss: 0.647221 valid's binary_logloss: 0.651249 [600] train's binary_logloss: 0.646154 valid's binary_logloss: 0.650938 [700] train's binary_logloss: 0.645276 valid's binary_logloss: 0.650398 [800] train's binary_logloss: 0.644511 valid's binary_logloss: 0.65025 [900] train's binary_logloss: 0.643837 valid's binary_logloss: 0.650107 [1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721 regularization_factors, val_score: 0.649721: 40%|#### | 8/20 [00:11<00:17, 1.49s/it][I 2020-09-27 04:44:54,488] Trial 47 finished with value: 0.6497208048936114 and parameters: {'lambda_l1': 6.216594964021316e-06, 'lambda_l2': 0.0003007341715934148}. Best is trial 45 with value: 0.6497207058626523. regularization_factors, val_score: 0.649721: 40%|#### | 8/20 [00:11<00:17, 1.49s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000538 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844 [200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602 [300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818 [400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001 [500] train's binary_logloss: 0.647226 valid's binary_logloss: 0.65118 [600] train's binary_logloss: 0.64616 valid's binary_logloss: 0.650744 [700] train's binary_logloss: 0.645279 valid's binary_logloss: 0.650352 [800] train's binary_logloss: 0.644543 valid's binary_logloss: 0.650069 [900] train's binary_logloss: 0.643859 valid's binary_logloss: 0.649948 [1000] train's binary_logloss: 0.643284 valid's binary_logloss: 0.649627 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.643284 valid's binary_logloss: 0.649627 regularization_factors, val_score: 0.649627: 45%|####5 | 9/20 [00:13<00:15, 1.45s/it][I 2020-09-27 04:44:55,828] Trial 48 finished with value: 0.6496273811877121 and parameters: {'lambda_l1': 0.0008017384871250849, 'lambda_l2': 2.2446245169734426e-07}. Best is trial 48 with value: 0.6496273811877121. regularization_factors, val_score: 0.649627: 45%|####5 | 9/20 [00:13<00:15, 1.45s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000443 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667565 valid's binary_logloss: 0.669845 [200] train's binary_logloss: 0.657208 valid's binary_logloss: 0.660603 [300] train's binary_logloss: 0.651961 valid's binary_logloss: 0.65582 [400] train's binary_logloss: 0.648962 valid's binary_logloss: 0.65306 [500] train's binary_logloss: 0.647204 valid's binary_logloss: 0.651318 [600] train's binary_logloss: 0.646126 valid's binary_logloss: 0.650793 [700] train's binary_logloss: 0.645262 valid's binary_logloss: 0.650639 [800] train's binary_logloss: 0.644495 valid's binary_logloss: 0.650417 [900] train's binary_logloss: 0.643822 valid's binary_logloss: 0.650127 [1000] train's binary_logloss: 0.643242 valid's binary_logloss: 0.649774 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.643242 valid's binary_logloss: 0.649774 regularization_factors, val_score: 0.649627: 50%|##### | 10/20 [00:15<00:15, 1.53s/it][I 2020-09-27 04:44:57,545] Trial 49 finished with value: 0.6497744578765494 and parameters: {'lambda_l1': 0.012286863044744575, 'lambda_l2': 1.2600943736007061e-08}. Best is trial 48 with value: 0.6496273811877121. regularization_factors, val_score: 0.649627: 50%|##### | 10/20 [00:15<00:15, 1.53s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000454 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667583 valid's binary_logloss: 0.6698 [200] train's binary_logloss: 0.657274 valid's binary_logloss: 0.66074 [300] train's binary_logloss: 0.652046 valid's binary_logloss: 0.655766 [400] train's binary_logloss: 0.649051 valid's binary_logloss: 0.65315 [500] train's binary_logloss: 0.647312 valid's binary_logloss: 0.651523 [600] train's binary_logloss: 0.646232 valid's binary_logloss: 0.651124 [700] train's binary_logloss: 0.645397 valid's binary_logloss: 0.650768 [800] train's binary_logloss: 0.644668 valid's binary_logloss: 0.650466 [900] train's binary_logloss: 0.643987 valid's binary_logloss: 0.650261 [1000] train's binary_logloss: 0.643431 valid's binary_logloss: 0.649835 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.643431 valid's binary_logloss: 0.649835 regularization_factors, val_score: 0.649627: 55%|#####5 | 11/20 [00:16<00:13, 1.47s/it][I 2020-09-27 04:44:58,890] Trial 50 finished with value: 0.6498345397919513 and parameters: {'lambda_l1': 0.0005572386578667407, 'lambda_l2': 2.4208863399594898}. Best is trial 48 with value: 0.6496273811877121. regularization_factors, val_score: 0.649627: 55%|#####5 | 11/20 [00:16<00:13, 1.47s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000464 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844 [200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602 [300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818 [400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001 [500] train's binary_logloss: 0.647221 valid's binary_logloss: 0.651249 [600] train's binary_logloss: 0.646154 valid's binary_logloss: 0.650938 [700] train's binary_logloss: 0.645276 valid's binary_logloss: 0.650398 [800] train's binary_logloss: 0.644511 valid's binary_logloss: 0.65025 [900] train's binary_logloss: 0.643837 valid's binary_logloss: 0.650107 [1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721 regularization_factors, val_score: 0.649627: 60%|###### | 12/20 [00:18<00:12, 1.52s/it][I 2020-09-27 04:45:00,537] Trial 51 finished with value: 0.6497208015259045 and parameters: {'lambda_l1': 8.699166692204462e-05, 'lambda_l2': 3.087574877344173e-06}. Best is trial 48 with value: 0.6496273811877121. regularization_factors, val_score: 0.649627: 60%|###### | 12/20 [00:18<00:12, 1.52s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002115 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844 [200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602 [300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818 [400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001 [500] train's binary_logloss: 0.647221 valid's binary_logloss: 0.651249 [600] train's binary_logloss: 0.646154 valid's binary_logloss: 0.650938 [700] train's binary_logloss: 0.645276 valid's binary_logloss: 0.650398 [800] train's binary_logloss: 0.644511 valid's binary_logloss: 0.65025 [900] train's binary_logloss: 0.643837 valid's binary_logloss: 0.650107 [1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721 regularization_factors, val_score: 0.649627: 65%|######5 | 13/20 [00:19<00:10, 1.50s/it][I 2020-09-27 04:45:01,964] Trial 52 finished with value: 0.6497208051737818 and parameters: {'lambda_l1': 0.00018230426317109586, 'lambda_l2': 4.548397960595403e-07}. Best is trial 48 with value: 0.6496273811877121. regularization_factors, val_score: 0.649627: 65%|######5 | 13/20 [00:19<00:10, 1.50s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000506 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844 [200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602 [300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818 [400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001 [500] train's binary_logloss: 0.647221 valid's binary_logloss: 0.651249 [600] train's binary_logloss: 0.646154 valid's binary_logloss: 0.650938 [700] train's binary_logloss: 0.645276 valid's binary_logloss: 0.650398 [800] train's binary_logloss: 0.644511 valid's binary_logloss: 0.65025 [900] train's binary_logloss: 0.643837 valid's binary_logloss: 0.650107 [1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721 regularization_factors, val_score: 0.649627: 70%|####### | 14/20 [00:20<00:08, 1.43s/it][I 2020-09-27 04:45:03,256] Trial 53 finished with value: 0.6497208043193139 and parameters: {'lambda_l1': 0.00015987949714084808, 'lambda_l2': 1.645060058121504e-06}. Best is trial 48 with value: 0.6496273811877121. regularization_factors, val_score: 0.649627: 70%|####### | 14/20 [00:20<00:08, 1.43s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000478 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844 [200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602 [300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818 [400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001 [500] train's binary_logloss: 0.647221 valid's binary_logloss: 0.651249 [600] train's binary_logloss: 0.646154 valid's binary_logloss: 0.650938 [700] train's binary_logloss: 0.645276 valid's binary_logloss: 0.650398 [800] train's binary_logloss: 0.644511 valid's binary_logloss: 0.650251 [900] train's binary_logloss: 0.643837 valid's binary_logloss: 0.650107 [1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721 regularization_factors, val_score: 0.649627: 75%|#######5 | 15/20 [00:22<00:07, 1.54s/it][I 2020-09-27 04:45:05,051] Trial 54 finished with value: 0.6497208216187388 and parameters: {'lambda_l1': 0.0006042412420761168, 'lambda_l2': 1.2746381508914564e-06}. Best is trial 48 with value: 0.6496273811877121. regularization_factors, val_score: 0.649627: 75%|#######5 | 15/20 [00:22<00:07, 1.54s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000393 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844 [200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602 [300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818 [400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001 [500] train's binary_logloss: 0.647221 valid's binary_logloss: 0.651249 [600] train's binary_logloss: 0.646154 valid's binary_logloss: 0.650938 [700] train's binary_logloss: 0.645276 valid's binary_logloss: 0.650398 [800] train's binary_logloss: 0.644511 valid's binary_logloss: 0.65025 [900] train's binary_logloss: 0.643837 valid's binary_logloss: 0.650107 [1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721 regularization_factors, val_score: 0.649627: 80%|######## | 16/20 [00:23<00:05, 1.50s/it][I 2020-09-27 04:45:06,437] Trial 55 finished with value: 0.6497208030993801 and parameters: {'lambda_l1': 0.00012765660632242594, 'lambda_l2': 3.053198787844645e-06}. Best is trial 48 with value: 0.6496273811877121. regularization_factors, val_score: 0.649627: 80%|######## | 16/20 [00:23<00:05, 1.50s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.005265 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844 [200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602 [300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818 [400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001 [500] train's binary_logloss: 0.647221 valid's binary_logloss: 0.651249 [600] train's binary_logloss: 0.646154 valid's binary_logloss: 0.650938 [700] train's binary_logloss: 0.645276 valid's binary_logloss: 0.650398 [800] train's binary_logloss: 0.644511 valid's binary_logloss: 0.65025 [900] train's binary_logloss: 0.643837 valid's binary_logloss: 0.650107 [1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.643265 valid's binary_logloss: 0.649721 regularization_factors, val_score: 0.649627: 85%|########5 | 17/20 [00:25<00:04, 1.47s/it][I 2020-09-27 04:45:07,858] Trial 56 finished with value: 0.6497208043349164 and parameters: {'lambda_l1': 4.556413554561587e-05, 'lambda_l2': 1.5942089823943423e-05}. Best is trial 48 with value: 0.6496273811877121. regularization_factors, val_score: 0.649627: 85%|########5 | 17/20 [00:25<00:04, 1.47s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000436 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667565 valid's binary_logloss: 0.669845 [200] train's binary_logloss: 0.657208 valid's binary_logloss: 0.660602 [300] train's binary_logloss: 0.651961 valid's binary_logloss: 0.655819 [400] train's binary_logloss: 0.648963 valid's binary_logloss: 0.65306 [500] train's binary_logloss: 0.647204 valid's binary_logloss: 0.651318 [600] train's binary_logloss: 0.646126 valid's binary_logloss: 0.650793 [700] train's binary_logloss: 0.645263 valid's binary_logloss: 0.650638 [800] train's binary_logloss: 0.644496 valid's binary_logloss: 0.650417 [900] train's binary_logloss: 0.643824 valid's binary_logloss: 0.650127 [1000] train's binary_logloss: 0.643243 valid's binary_logloss: 0.649775 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.643243 valid's binary_logloss: 0.649775 regularization_factors, val_score: 0.649627: 90%|######### | 18/20 [00:26<00:03, 1.52s/it][I 2020-09-27 04:45:09,475] Trial 57 finished with value: 0.6497749853707365 and parameters: {'lambda_l1': 9.790223265016275e-08, 'lambda_l2': 0.04135439527004478}. Best is trial 48 with value: 0.6496273811877121. regularization_factors, val_score: 0.649627: 90%|######### | 18/20 [00:26<00:03, 1.52s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000415 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669845 [200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602 [300] train's binary_logloss: 0.65196 valid's binary_logloss: 0.655819 [400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001 [500] train's binary_logloss: 0.647227 valid's binary_logloss: 0.65118 [600] train's binary_logloss: 0.64616 valid's binary_logloss: 0.650744 [700] train's binary_logloss: 0.64528 valid's binary_logloss: 0.650352 [800] train's binary_logloss: 0.644543 valid's binary_logloss: 0.65007 [900] train's binary_logloss: 0.64386 valid's binary_logloss: 0.649948 [1000] train's binary_logloss: 0.643285 valid's binary_logloss: 0.649628 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.643285 valid's binary_logloss: 0.649628 regularization_factors, val_score: 0.649627: 95%|#########5| 19/20 [00:28<00:01, 1.46s/it][I 2020-09-27 04:45:10,790] Trial 58 finished with value: 0.6496276632112322 and parameters: {'lambda_l1': 0.0038892159453483817, 'lambda_l2': 3.162000398193428e-08}. Best is trial 48 with value: 0.6496273811877121. regularization_factors, val_score: 0.649627: 95%|#########5| 19/20 [00:28<00:01, 1.46s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000477 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667565 valid's binary_logloss: 0.669846 [200] train's binary_logloss: 0.657208 valid's binary_logloss: 0.660603 [300] train's binary_logloss: 0.651961 valid's binary_logloss: 0.65582 [400] train's binary_logloss: 0.648963 valid's binary_logloss: 0.65306 [500] train's binary_logloss: 0.647204 valid's binary_logloss: 0.651318 [600] train's binary_logloss: 0.646126 valid's binary_logloss: 0.650793 [700] train's binary_logloss: 0.645262 valid's binary_logloss: 0.650639 [800] train's binary_logloss: 0.644495 valid's binary_logloss: 0.650418 [900] train's binary_logloss: 0.643823 valid's binary_logloss: 0.650127 [1000] train's binary_logloss: 0.643243 valid's binary_logloss: 0.649775 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.643243 valid's binary_logloss: 0.649775 regularization_factors, val_score: 0.649627: 100%|##########| 20/20 [00:30<00:00, 1.59s/it][I 2020-09-27 04:45:12,683] Trial 59 finished with value: 0.6497745906504417 and parameters: {'lambda_l1': 0.012927722016684172, 'lambda_l2': 1.3099842616085507e-08}. Best is trial 48 with value: 0.6496273811877121. regularization_factors, val_score: 0.649627: 100%|##########| 20/20 [00:30<00:00, 1.51s/it] min_data_in_leaf, val_score: 0.649627: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000494 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844 [200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602 [300] train's binary_logloss: 0.652009 valid's binary_logloss: 0.655779 [400] train's binary_logloss: 0.649064 valid's binary_logloss: 0.653022 [500] train's binary_logloss: 0.647474 valid's binary_logloss: 0.651157 [600] train's binary_logloss: 0.646465 valid's binary_logloss: 0.650811 [700] train's binary_logloss: 0.645712 valid's binary_logloss: 0.650417 [800] train's binary_logloss: 0.645041 valid's binary_logloss: 0.650168 [900] train's binary_logloss: 0.644428 valid's binary_logloss: 0.650337 Early stopping, best iteration is: [811] train's binary_logloss: 0.644942 valid's binary_logloss: 0.65 min_data_in_leaf, val_score: 0.649627: 20%|## | 1/5 [00:01<00:05, 1.26s/it][I 2020-09-27 04:45:13,958] Trial 60 finished with value: 0.6500004379994676 and parameters: {'min_child_samples': 100}. Best is trial 60 with value: 0.6500004379994676. min_data_in_leaf, val_score: 0.649627: 20%|## | 1/5 [00:01<00:05, 1.26s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000400 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844 [200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602 [300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818 [400] train's binary_logloss: 0.648932 valid's binary_logloss: 0.652828 [500] train's binary_logloss: 0.647269 valid's binary_logloss: 0.651176 [600] train's binary_logloss: 0.646216 valid's binary_logloss: 0.650912 [700] train's binary_logloss: 0.64537 valid's binary_logloss: 0.650588 [800] train's binary_logloss: 0.644654 valid's binary_logloss: 0.65043 [900] train's binary_logloss: 0.643973 valid's binary_logloss: 0.650308 Early stopping, best iteration is: [811] train's binary_logloss: 0.644548 valid's binary_logloss: 0.650197 min_data_in_leaf, val_score: 0.649627: 40%|#### | 2/5 [00:02<00:03, 1.26s/it][I 2020-09-27 04:45:15,217] Trial 61 finished with value: 0.6501970617141319 and parameters: {'min_child_samples': 50}. Best is trial 60 with value: 0.6500004379994676. min_data_in_leaf, val_score: 0.649627: 40%|#### | 2/5 [00:02<00:03, 1.26s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000404 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844 [200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602 [300] train's binary_logloss: 0.65195 valid's binary_logloss: 0.65587 [400] train's binary_logloss: 0.648925 valid's binary_logloss: 0.65304 [500] train's binary_logloss: 0.647138 valid's binary_logloss: 0.651221 [600] train's binary_logloss: 0.646029 valid's binary_logloss: 0.650864 [700] train's binary_logloss: 0.645079 valid's binary_logloss: 0.650534 [800] train's binary_logloss: 0.644313 valid's binary_logloss: 0.650116 [900] train's binary_logloss: 0.643631 valid's binary_logloss: 0.650098 Early stopping, best iteration is: [811] train's binary_logloss: 0.644204 valid's binary_logloss: 0.649909 min_data_in_leaf, val_score: 0.649627: 60%|###### | 3/5 [00:04<00:02, 1.36s/it][I 2020-09-27 04:45:16,819] Trial 62 finished with value: 0.6499094175863028 and parameters: {'min_child_samples': 5}. Best is trial 62 with value: 0.6499094175863028. min_data_in_leaf, val_score: 0.649627: 60%|###### | 3/5 [00:04<00:02, 1.36s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000374 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844 [200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602 [300] train's binary_logloss: 0.65195 valid's binary_logloss: 0.65587 [400] train's binary_logloss: 0.648927 valid's binary_logloss: 0.653037 [500] train's binary_logloss: 0.647134 valid's binary_logloss: 0.651144 [600] train's binary_logloss: 0.646045 valid's binary_logloss: 0.650746 [700] train's binary_logloss: 0.64509 valid's binary_logloss: 0.650337 [800] train's binary_logloss: 0.644337 valid's binary_logloss: 0.650057 [900] train's binary_logloss: 0.643666 valid's binary_logloss: 0.649979 Early stopping, best iteration is: [811] train's binary_logloss: 0.64422 valid's binary_logloss: 0.649764 min_data_in_leaf, val_score: 0.649627: 80%|######## | 4/5 [00:05<00:01, 1.32s/it][I 2020-09-27 04:45:18,047] Trial 63 finished with value: 0.649763563681734 and parameters: {'min_child_samples': 10}. Best is trial 63 with value: 0.649763563681734. min_data_in_leaf, val_score: 0.649627: 80%|######## | 4/5 [00:05<00:01, 1.32s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000400 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667564 valid's binary_logloss: 0.669844 [200] train's binary_logloss: 0.657207 valid's binary_logloss: 0.660602 [300] train's binary_logloss: 0.651959 valid's binary_logloss: 0.655818 [400] train's binary_logloss: 0.648965 valid's binary_logloss: 0.653001 [500] train's binary_logloss: 0.647226 valid's binary_logloss: 0.65118 [600] train's binary_logloss: 0.646171 valid's binary_logloss: 0.650777 [700] train's binary_logloss: 0.645308 valid's binary_logloss: 0.650525 [800] train's binary_logloss: 0.644563 valid's binary_logloss: 0.650372 [900] train's binary_logloss: 0.643935 valid's binary_logloss: 0.650147 [1000] train's binary_logloss: 0.643333 valid's binary_logloss: 0.649925 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.643333 valid's binary_logloss: 0.649925 min_data_in_leaf, val_score: 0.649627: 100%|##########| 5/5 [00:06<00:00, 1.32s/it][I 2020-09-27 04:45:19,357] Trial 64 finished with value: 0.6499246052491834 and parameters: {'min_child_samples': 25}. Best is trial 63 with value: 0.649763563681734. min_data_in_leaf, val_score: 0.649627: 100%|##########| 5/5 [00:06<00:00, 1.33s/it] Fold : 7 [I 2020-09-27 04:45:19,501] A new study created in memory with name: no-name-a7907594-6487-49b5-9a75-d6ef7c2f3b45 feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008857 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.576754 valid's binary_logloss: 0.65897 Early stopping, best iteration is: [69] train's binary_logloss: 0.59731 valid's binary_logloss: 0.657355 feature_fraction, val_score: 0.657355: 14%|#4 | 1/7 [00:00<00:05, 1.05it/s][I 2020-09-27 04:45:20,471] Trial 0 finished with value: 0.6573548252031896 and parameters: {'feature_fraction': 0.6}. Best is trial 0 with value: 0.6573548252031896. feature_fraction, val_score: 0.657355: 14%|#4 | 1/7 [00:00<00:05, 1.05it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000492 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.578741 valid's binary_logloss: 0.657569 [200] train's binary_logloss: 0.527428 valid's binary_logloss: 0.66217 Early stopping, best iteration is: [101] train's binary_logloss: 0.578076 valid's binary_logloss: 0.657517 feature_fraction, val_score: 0.657355: 29%|##8 | 2/7 [00:01<00:04, 1.19it/s][I 2020-09-27 04:45:21,053] Trial 1 finished with value: 0.6575168666853799 and parameters: {'feature_fraction': 0.5}. Best is trial 0 with value: 0.6573548252031896. feature_fraction, val_score: 0.657355: 29%|##8 | 2/7 [00:01<00:04, 1.19it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000966 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.572623 valid's binary_logloss: 0.65727 Early stopping, best iteration is: [81] train's binary_logloss: 0.585015 valid's binary_logloss: 0.656834 feature_fraction, val_score: 0.656834: 43%|####2 | 3/7 [00:02<00:03, 1.31it/s][I 2020-09-27 04:45:21,630] Trial 2 finished with value: 0.6568343510931819 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 2 with value: 0.6568343510931819. feature_fraction, val_score: 0.656834: 43%|####2 | 3/7 [00:02<00:03, 1.31it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008151 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.573507 valid's binary_logloss: 0.657017 [200] train's binary_logloss: 0.517816 valid's binary_logloss: 0.65931 Early stopping, best iteration is: [131] train's binary_logloss: 0.55462 valid's binary_logloss: 0.655904 feature_fraction, val_score: 0.655904: 57%|#####7 | 4/7 [00:02<00:02, 1.36it/s][I 2020-09-27 04:45:22,306] Trial 3 finished with value: 0.6559041635016774 and parameters: {'feature_fraction': 0.8}. Best is trial 3 with value: 0.6559041635016774. feature_fraction, val_score: 0.655904: 57%|#####7 | 4/7 [00:02<00:02, 1.36it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005133 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.576598 valid's binary_logloss: 0.657946 Early stopping, best iteration is: [91] train's binary_logloss: 0.582318 valid's binary_logloss: 0.656196 feature_fraction, val_score: 0.655904: 71%|#######1 | 5/7 [00:03<00:01, 1.49it/s][I 2020-09-27 04:45:22,831] Trial 4 finished with value: 0.6561957860921139 and parameters: {'feature_fraction': 0.7}. Best is trial 3 with value: 0.6559041635016774. feature_fraction, val_score: 0.655904: 71%|#######1 | 5/7 [00:03<00:01, 1.49it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000955 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.570187 valid's binary_logloss: 0.658113 Early stopping, best iteration is: [86] train's binary_logloss: 0.579352 valid's binary_logloss: 0.657006 feature_fraction, val_score: 0.655904: 86%|########5 | 6/7 [00:04<00:00, 1.27it/s][I 2020-09-27 04:45:23,895] Trial 5 finished with value: 0.6570061744677883 and parameters: {'feature_fraction': 1.0}. Best is trial 3 with value: 0.6559041635016774. feature_fraction, val_score: 0.655904: 86%|########5 | 6/7 [00:04<00:00, 1.27it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000543 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.582693 valid's binary_logloss: 0.657081 [200] train's binary_logloss: 0.532209 valid's binary_logloss: 0.662404 Early stopping, best iteration is: [100] train's binary_logloss: 0.582693 valid's binary_logloss: 0.657081 feature_fraction, val_score: 0.655904: 100%|##########| 7/7 [00:04<00:00, 1.39it/s][I 2020-09-27 04:45:24,444] Trial 6 finished with value: 0.6570805680910938 and parameters: {'feature_fraction': 0.4}. Best is trial 3 with value: 0.6559041635016774. feature_fraction, val_score: 0.655904: 100%|##########| 7/7 [00:04<00:00, 1.42it/s] num_leaves, val_score: 0.655904: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004860 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.466323 valid's binary_logloss: 0.658116 Early stopping, best iteration is: [54] train's binary_logloss: 0.534041 valid's binary_logloss: 0.655259 num_leaves, val_score: 0.655259: 5%|5 | 1/20 [00:00<00:14, 1.35it/s][I 2020-09-27 04:45:25,204] Trial 7 finished with value: 0.6552590163821485 and parameters: {'num_leaves': 85}. Best is trial 7 with value: 0.6552590163821485. num_leaves, val_score: 0.655259: 5%|5 | 1/20 [00:00<00:14, 1.35it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000895 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.540475 valid's binary_logloss: 0.662426 Early stopping, best iteration is: [83] train's binary_logloss: 0.556094 valid's binary_logloss: 0.65974 num_leaves, val_score: 0.655259: 10%|# | 2/20 [00:01<00:12, 1.40it/s][I 2020-09-27 04:45:25,845] Trial 8 finished with value: 0.6597395208090874 and parameters: {'num_leaves': 46}. Best is trial 7 with value: 0.6552590163821485. num_leaves, val_score: 0.655259: 10%|# | 2/20 [00:01<00:12, 1.40it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004966 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.60271 valid's binary_logloss: 0.656639 [200] train's binary_logloss: 0.564039 valid's binary_logloss: 0.658129 Early stopping, best iteration is: [108] train's binary_logloss: 0.59936 valid's binary_logloss: 0.65622 num_leaves, val_score: 0.655259: 15%|#5 | 3/20 [00:01<00:11, 1.53it/s][I 2020-09-27 04:45:26,358] Trial 9 finished with value: 0.6562201154063112 and parameters: {'num_leaves': 19}. Best is trial 7 with value: 0.6552590163821485. num_leaves, val_score: 0.655259: 15%|#5 | 3/20 [00:01<00:11, 1.53it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000795 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.314803 valid's binary_logloss: 0.675418 Early stopping, best iteration is: [17] train's binary_logloss: 0.568221 valid's binary_logloss: 0.663895 num_leaves, val_score: 0.655259: 20%|## | 4/20 [00:03<00:15, 1.05it/s][I 2020-09-27 04:45:28,019] Trial 10 finished with value: 0.6638954298043613 and parameters: {'num_leaves': 201}. Best is trial 7 with value: 0.6552590163821485. num_leaves, val_score: 0.655259: 20%|## | 4/20 [00:03<00:15, 1.05it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001158 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.415347 valid's binary_logloss: 0.669083 Early stopping, best iteration is: [25] train's binary_logloss: 0.574128 valid's binary_logloss: 0.660544 num_leaves, val_score: 0.655259: 25%|##5 | 5/20 [00:04<00:13, 1.10it/s][I 2020-09-27 04:45:28,824] Trial 11 finished with value: 0.6605435697983096 and parameters: {'num_leaves': 118}. Best is trial 7 with value: 0.6552590163821485. num_leaves, val_score: 0.655259: 25%|##5 | 5/20 [00:04<00:13, 1.10it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000788 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.424268 valid's binary_logloss: 0.667432 Early stopping, best iteration is: [24] train's binary_logloss: 0.581498 valid's binary_logloss: 0.659094 num_leaves, val_score: 0.655259: 30%|### | 6/20 [00:05<00:12, 1.16it/s][I 2020-09-27 04:45:29,584] Trial 12 finished with value: 0.6590937806617987 and parameters: {'num_leaves': 112}. Best is trial 7 with value: 0.6552590163821485. num_leaves, val_score: 0.655259: 30%|### | 6/20 [00:05<00:12, 1.16it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000922 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.264846 valid's binary_logloss: 0.665644 Early stopping, best iteration is: [38] train's binary_logloss: 0.441545 valid's binary_logloss: 0.654899 num_leaves, val_score: 0.654899: 35%|###5 | 7/20 [00:07<00:15, 1.17s/it][I 2020-09-27 04:45:31,476] Trial 13 finished with value: 0.6548985372701535 and parameters: {'num_leaves': 256}. Best is trial 13 with value: 0.6548985372701535. num_leaves, val_score: 0.654899: 35%|###5 | 7/20 [00:07<00:15, 1.17s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000853 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.276152 valid's binary_logloss: 0.674007 Early stopping, best iteration is: [19] train's binary_logloss: 0.540876 valid's binary_logloss: 0.662293 num_leaves, val_score: 0.654899: 40%|#### | 8/20 [00:08<00:14, 1.17s/it][I 2020-09-27 04:45:32,652] Trial 14 finished with value: 0.6622934249190215 and parameters: {'num_leaves': 244}. Best is trial 13 with value: 0.6548985372701535. num_leaves, val_score: 0.654899: 40%|#### | 8/20 [00:08<00:14, 1.17s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000843 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.475337 valid's binary_logloss: 0.661813 Early stopping, best iteration is: [30] train's binary_logloss: 0.586366 valid's binary_logloss: 0.658471 num_leaves, val_score: 0.654899: 45%|####5 | 9/20 [00:08<00:11, 1.02s/it][I 2020-09-27 04:45:33,298] Trial 15 finished with value: 0.6584705405485446 and parameters: {'num_leaves': 81}. Best is trial 13 with value: 0.6548985372701535. num_leaves, val_score: 0.654899: 45%|####5 | 9/20 [00:08<00:11, 1.02s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009872 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.32841 valid's binary_logloss: 0.676457 Early stopping, best iteration is: [23] train's binary_logloss: 0.544414 valid's binary_logloss: 0.664223 num_leaves, val_score: 0.654899: 50%|##### | 10/20 [00:09<00:09, 1.01it/s][I 2020-09-27 04:45:34,239] Trial 16 finished with value: 0.664222940902035 and parameters: {'num_leaves': 189}. Best is trial 13 with value: 0.6548985372701535. num_leaves, val_score: 0.654899: 50%|##### | 10/20 [00:09<00:09, 1.01it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010397 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.353294 valid's binary_logloss: 0.668112 Early stopping, best iteration is: [38] train's binary_logloss: 0.499593 valid's binary_logloss: 0.659529 num_leaves, val_score: 0.654899: 55%|#####5 | 11/20 [00:11<00:10, 1.17s/it][I 2020-09-27 04:45:35,832] Trial 17 finished with value: 0.6595289898494625 and parameters: {'num_leaves': 166}. Best is trial 13 with value: 0.6548985372701535. num_leaves, val_score: 0.654899: 55%|#####5 | 11/20 [00:11<00:10, 1.17s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000949 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.263628 valid's binary_logloss: 0.683347 Early stopping, best iteration is: [22] train's binary_logloss: 0.520072 valid's binary_logloss: 0.662143 num_leaves, val_score: 0.654899: 60%|###### | 12/20 [00:12<00:09, 1.22s/it][I 2020-09-27 04:45:37,170] Trial 18 finished with value: 0.6621427253155968 and parameters: {'num_leaves': 254}. Best is trial 13 with value: 0.6548985372701535. num_leaves, val_score: 0.654899: 60%|###### | 12/20 [00:12<00:09, 1.22s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004879 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.505464 valid's binary_logloss: 0.660159 Early stopping, best iteration is: [34] train's binary_logloss: 0.592348 valid's binary_logloss: 0.657805 num_leaves, val_score: 0.654899: 65%|######5 | 13/20 [00:13<00:07, 1.02s/it][I 2020-09-27 04:45:37,726] Trial 19 finished with value: 0.6578051021077662 and parameters: {'num_leaves': 64}. Best is trial 13 with value: 0.6548985372701535. num_leaves, val_score: 0.654899: 65%|######5 | 13/20 [00:13<00:07, 1.02s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000866 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.373419 valid's binary_logloss: 0.673149 Early stopping, best iteration is: [25] train's binary_logloss: 0.55608 valid's binary_logloss: 0.664823 num_leaves, val_score: 0.654899: 70%|####### | 14/20 [00:14<00:06, 1.13s/it][I 2020-09-27 04:45:39,108] Trial 20 finished with value: 0.6648226497863418 and parameters: {'num_leaves': 150}. Best is trial 13 with value: 0.6548985372701535. num_leaves, val_score: 0.654899: 70%|####### | 14/20 [00:14<00:06, 1.13s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000806 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.447378 valid's binary_logloss: 0.661981 Early stopping, best iteration is: [54] train's binary_logloss: 0.521425 valid's binary_logloss: 0.658385 num_leaves, val_score: 0.654899: 75%|#######5 | 15/20 [00:15<00:05, 1.04s/it][I 2020-09-27 04:45:39,938] Trial 21 finished with value: 0.6583849556869334 and parameters: {'num_leaves': 97}. Best is trial 13 with value: 0.6548985372701535. num_leaves, val_score: 0.654899: 75%|#######5 | 15/20 [00:15<00:05, 1.04s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004821 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650824 valid's binary_logloss: 0.661115 [200] train's binary_logloss: 0.639749 valid's binary_logloss: 0.657461 [300] train's binary_logloss: 0.632853 valid's binary_logloss: 0.656219 Early stopping, best iteration is: [286] train's binary_logloss: 0.633724 valid's binary_logloss: 0.656027 num_leaves, val_score: 0.654899: 80%|######## | 16/20 [00:16<00:03, 1.08it/s][I 2020-09-27 04:45:40,584] Trial 22 finished with value: 0.6560272050011768 and parameters: {'num_leaves': 4}. Best is trial 13 with value: 0.6548985372701535. num_leaves, val_score: 0.654899: 80%|######## | 16/20 [00:16<00:03, 1.08it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000847 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.295072 valid's binary_logloss: 0.672517 Early stopping, best iteration is: [35] train's binary_logloss: 0.474148 valid's binary_logloss: 0.65871 num_leaves, val_score: 0.654899: 85%|########5 | 17/20 [00:17<00:03, 1.02s/it][I 2020-09-27 04:45:41,847] Trial 23 finished with value: 0.6587097853479928 and parameters: {'num_leaves': 222}. Best is trial 13 with value: 0.6548985372701535. num_leaves, val_score: 0.654899: 85%|########5 | 17/20 [00:17<00:03, 1.02s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.006283 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.559059 valid's binary_logloss: 0.657941 Early stopping, best iteration is: [78] train's binary_logloss: 0.57617 valid's binary_logloss: 0.656731 num_leaves, val_score: 0.654899: 90%|######### | 18/20 [00:18<00:02, 1.04s/it][I 2020-09-27 04:45:42,913] Trial 24 finished with value: 0.6567308128514865 and parameters: {'num_leaves': 37}. Best is trial 13 with value: 0.6548985372701535. num_leaves, val_score: 0.654899: 90%|######### | 18/20 [00:18<00:02, 1.04s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010220 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.388417 valid's binary_logloss: 0.668977 Early stopping, best iteration is: [38] train's binary_logloss: 0.521183 valid's binary_logloss: 0.659883 num_leaves, val_score: 0.654899: 95%|#########5| 19/20 [00:19<00:01, 1.02s/it][I 2020-09-27 04:45:43,893] Trial 25 finished with value: 0.659882543351282 and parameters: {'num_leaves': 138}. Best is trial 13 with value: 0.6548985372701535. num_leaves, val_score: 0.654899: 95%|#########5| 19/20 [00:19<00:01, 1.02s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001162 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.465903 valid's binary_logloss: 0.661386 Early stopping, best iteration is: [58] train's binary_logloss: 0.526905 valid's binary_logloss: 0.657041 num_leaves, val_score: 0.654899: 100%|##########| 20/20 [00:20<00:00, 1.03it/s][I 2020-09-27 04:45:44,739] Trial 26 finished with value: 0.6570410039446851 and parameters: {'num_leaves': 86}. Best is trial 13 with value: 0.6548985372701535. num_leaves, val_score: 0.654899: 100%|##########| 20/20 [00:20<00:00, 1.01s/it] bagging, val_score: 0.654899: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004721 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.278384 valid's binary_logloss: 0.695429 Early stopping, best iteration is: [15] train's binary_logloss: 0.574718 valid's binary_logloss: 0.670457 bagging, val_score: 0.654899: 10%|# | 1/10 [00:01<00:10, 1.20s/it][I 2020-09-27 04:45:45,957] Trial 27 finished with value: 0.6704565961142148 and parameters: {'bagging_fraction': 0.5211974818467775, 'bagging_freq': 5}. Best is trial 27 with value: 0.6704565961142148. bagging, val_score: 0.654899: 10%|# | 1/10 [00:01<00:10, 1.20s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.018210 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.262977 valid's binary_logloss: 0.676276 Early stopping, best iteration is: [30] train's binary_logloss: 0.4766 valid's binary_logloss: 0.659897 bagging, val_score: 0.654899: 20%|## | 2/10 [00:02<00:10, 1.37s/it][I 2020-09-27 04:45:47,718] Trial 28 finished with value: 0.6598967542506259 and parameters: {'bagging_fraction': 0.9934815132316437, 'bagging_freq': 1}. Best is trial 28 with value: 0.6598967542506259. bagging, val_score: 0.654899: 20%|## | 2/10 [00:02<00:10, 1.37s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000944 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.264681 valid's binary_logloss: 0.684914 Early stopping, best iteration is: [17] train's binary_logloss: 0.549363 valid's binary_logloss: 0.665808 bagging, val_score: 0.654899: 30%|### | 3/10 [00:04<00:09, 1.35s/it][I 2020-09-27 04:45:49,025] Trial 29 finished with value: 0.6658075550967804 and parameters: {'bagging_fraction': 0.9461366644242709, 'bagging_freq': 7}. Best is trial 28 with value: 0.6598967542506259. bagging, val_score: 0.654899: 30%|### | 3/10 [00:04<00:09, 1.35s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000798 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.281086 valid's binary_logloss: 0.707672 Early stopping, best iteration is: [17] train's binary_logloss: 0.564424 valid's binary_logloss: 0.669487 bagging, val_score: 0.654899: 40%|#### | 4/10 [00:06<00:08, 1.49s/it][I 2020-09-27 04:45:50,825] Trial 30 finished with value: 0.6694865333360387 and parameters: {'bagging_fraction': 0.4198139622048719, 'bagging_freq': 1}. Best is trial 28 with value: 0.6598967542506259. bagging, val_score: 0.654899: 40%|#### | 4/10 [00:06<00:08, 1.49s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000847 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.266862 valid's binary_logloss: 0.689038 Early stopping, best iteration is: [24] train's binary_logloss: 0.514004 valid's binary_logloss: 0.66343 bagging, val_score: 0.654899: 50%|##### | 5/10 [00:07<00:07, 1.51s/it][I 2020-09-27 04:45:52,393] Trial 31 finished with value: 0.663429926936811 and parameters: {'bagging_fraction': 0.7512986978703702, 'bagging_freq': 4}. Best is trial 28 with value: 0.6598967542506259. bagging, val_score: 0.654899: 50%|##### | 5/10 [00:07<00:07, 1.51s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004195 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.265958 valid's binary_logloss: 0.68931 Early stopping, best iteration is: [18] train's binary_logloss: 0.549252 valid's binary_logloss: 0.6651 bagging, val_score: 0.654899: 60%|###### | 6/10 [00:09<00:05, 1.48s/it][I 2020-09-27 04:45:53,804] Trial 32 finished with value: 0.6650999977344688 and parameters: {'bagging_fraction': 0.7159980215103693, 'bagging_freq': 7}. Best is trial 28 with value: 0.6598967542506259. bagging, val_score: 0.654899: 60%|###### | 6/10 [00:09<00:05, 1.48s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.014957 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.272788 valid's binary_logloss: 0.699964 Early stopping, best iteration is: [22] train's binary_logloss: 0.532238 valid's binary_logloss: 0.667015 bagging, val_score: 0.654899: 70%|####### | 7/10 [00:10<00:04, 1.52s/it][I 2020-09-27 04:45:55,408] Trial 33 finished with value: 0.6670147726739027 and parameters: {'bagging_fraction': 0.5563678837629615, 'bagging_freq': 2}. Best is trial 28 with value: 0.6598967542506259. bagging, val_score: 0.654899: 70%|####### | 7/10 [00:10<00:04, 1.52s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000817 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.267946 valid's binary_logloss: 0.681235 Early stopping, best iteration is: [21] train's binary_logloss: 0.52815 valid's binary_logloss: 0.661059 bagging, val_score: 0.654899: 80%|######## | 8/10 [00:12<00:02, 1.47s/it][I 2020-09-27 04:45:56,765] Trial 34 finished with value: 0.6610593718244844 and parameters: {'bagging_fraction': 0.8313758530340752, 'bagging_freq': 3}. Best is trial 28 with value: 0.6598967542506259. bagging, val_score: 0.654899: 80%|######## | 8/10 [00:12<00:02, 1.47s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001483 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.292079 valid's binary_logloss: 0.717802 Early stopping, best iteration is: [16] train's binary_logloss: 0.575885 valid's binary_logloss: 0.667398 bagging, val_score: 0.654899: 90%|######### | 9/10 [00:13<00:01, 1.56s/it][I 2020-09-27 04:45:58,520] Trial 35 finished with value: 0.6673978773322778 and parameters: {'bagging_fraction': 0.4012246265638672, 'bagging_freq': 5}. Best is trial 28 with value: 0.6598967542506259. bagging, val_score: 0.654899: 90%|######### | 9/10 [00:13<00:01, 1.56s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.006489 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.274203 valid's binary_logloss: 0.693572 Early stopping, best iteration is: [23] train's binary_logloss: 0.525889 valid's binary_logloss: 0.667929 bagging, val_score: 0.654899: 100%|##########| 10/10 [00:14<00:00, 1.46s/it][I 2020-09-27 04:45:59,749] Trial 36 finished with value: 0.667928676492384 and parameters: {'bagging_fraction': 0.6154723571783186, 'bagging_freq': 6}. Best is trial 28 with value: 0.6598967542506259. bagging, val_score: 0.654899: 100%|##########| 10/10 [00:15<00:00, 1.50s/it] feature_fraction_stage2, val_score: 0.654899: 0%| | 0/6 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004865 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.269387 valid's binary_logloss: 0.674759 Early stopping, best iteration is: [25] train's binary_logloss: 0.505622 valid's binary_logloss: 0.660533 feature_fraction_stage2, val_score: 0.654899: 17%|#6 | 1/6 [00:01<00:05, 1.11s/it][I 2020-09-27 04:46:00,871] Trial 37 finished with value: 0.6605329410813362 and parameters: {'feature_fraction': 0.7200000000000001}. Best is trial 37 with value: 0.6605329410813362. feature_fraction_stage2, val_score: 0.654899: 17%|#6 | 1/6 [00:01<00:05, 1.11s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001084 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.264846 valid's binary_logloss: 0.665644 Early stopping, best iteration is: [38] train's binary_logloss: 0.441545 valid's binary_logloss: 0.654899 feature_fraction_stage2, val_score: 0.654899: 33%|###3 | 2/6 [00:03<00:05, 1.35s/it][I 2020-09-27 04:46:02,793] Trial 38 finished with value: 0.6548985372701535 and parameters: {'feature_fraction': 0.8160000000000001}. Best is trial 38 with value: 0.6548985372701535. feature_fraction_stage2, val_score: 0.654899: 33%|###3 | 2/6 [00:03<00:05, 1.35s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004430 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.261299 valid's binary_logloss: 0.680887 Early stopping, best iteration is: [20] train's binary_logloss: 0.526475 valid's binary_logloss: 0.664407 feature_fraction_stage2, val_score: 0.654899: 50%|##### | 3/6 [00:04<00:03, 1.29s/it][I 2020-09-27 04:46:03,948] Trial 39 finished with value: 0.6644073193077987 and parameters: {'feature_fraction': 0.88}. Best is trial 38 with value: 0.6548985372701535. feature_fraction_stage2, val_score: 0.654899: 50%|##### | 3/6 [00:04<00:03, 1.29s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000627 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.265461 valid's binary_logloss: 0.682897 Early stopping, best iteration is: [24] train's binary_logloss: 0.509383 valid's binary_logloss: 0.662438 feature_fraction_stage2, val_score: 0.654899: 67%|######6 | 4/6 [00:06<00:03, 1.74s/it][I 2020-09-27 04:46:06,737] Trial 40 finished with value: 0.6624379814652182 and parameters: {'feature_fraction': 0.7520000000000001}. Best is trial 38 with value: 0.6548985372701535. feature_fraction_stage2, val_score: 0.654899: 67%|######6 | 4/6 [00:06<00:03, 1.74s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005011 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.265461 valid's binary_logloss: 0.682897 Early stopping, best iteration is: [24] train's binary_logloss: 0.509383 valid's binary_logloss: 0.662438 feature_fraction_stage2, val_score: 0.654899: 83%|########3 | 5/6 [00:08<00:01, 1.57s/it][I 2020-09-27 04:46:07,913] Trial 41 finished with value: 0.6624379814652182 and parameters: {'feature_fraction': 0.784}. Best is trial 38 with value: 0.6548985372701535. feature_fraction_stage2, val_score: 0.654899: 83%|########3 | 5/6 [00:08<00:01, 1.57s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000490 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.261804 valid's binary_logloss: 0.683679 Early stopping, best iteration is: [25] train's binary_logloss: 0.500705 valid's binary_logloss: 0.666858 feature_fraction_stage2, val_score: 0.654899: 100%|##########| 6/6 [00:09<00:00, 1.64s/it][I 2020-09-27 04:46:09,700] Trial 42 finished with value: 0.6668584743708171 and parameters: {'feature_fraction': 0.8480000000000001}. Best is trial 38 with value: 0.6548985372701535. feature_fraction_stage2, val_score: 0.654899: 100%|##########| 6/6 [00:09<00:00, 1.66s/it] regularization_factors, val_score: 0.654899: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001030 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.265642 valid's binary_logloss: 0.678724 Early stopping, best iteration is: [32] train's binary_logloss: 0.469659 valid's binary_logloss: 0.660888 regularization_factors, val_score: 0.654899: 5%|5 | 1/20 [00:01<00:29, 1.58s/it][I 2020-09-27 04:46:11,296] Trial 43 finished with value: 0.6608876654466368 and parameters: {'lambda_l1': 2.395209311350816e-06, 'lambda_l2': 0.13259120292527007}. Best is trial 43 with value: 0.6608876654466368. regularization_factors, val_score: 0.654899: 5%|5 | 1/20 [00:01<00:29, 1.58s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000545 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.390048 valid's binary_logloss: 0.672992 Early stopping, best iteration is: [35] train's binary_logloss: 0.535479 valid's binary_logloss: 0.6619 regularization_factors, val_score: 0.654899: 10%|# | 2/20 [00:03<00:29, 1.62s/it][I 2020-09-27 04:46:13,025] Trial 44 finished with value: 0.6619000150383266 and parameters: {'lambda_l1': 6.3824374533786505, 'lambda_l2': 2.9167015559251105e-08}. Best is trial 43 with value: 0.6608876654466368. regularization_factors, val_score: 0.654899: 10%|# | 2/20 [00:03<00:29, 1.62s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000772 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.295559 valid's binary_logloss: 0.675205 Early stopping, best iteration is: [31] train's binary_logloss: 0.4927 valid's binary_logloss: 0.660235 regularization_factors, val_score: 0.654899: 15%|#5 | 3/20 [00:05<00:30, 1.78s/it][I 2020-09-27 04:46:15,164] Trial 45 finished with value: 0.6602349076536601 and parameters: {'lambda_l1': 1.6485218908617958, 'lambda_l2': 1.1064051625349032e-08}. Best is trial 45 with value: 0.6602349076536601. regularization_factors, val_score: 0.654899: 15%|#5 | 3/20 [00:05<00:30, 1.78s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000792 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.263724 valid's binary_logloss: 0.666125 Early stopping, best iteration is: [37] train's binary_logloss: 0.44514 valid's binary_logloss: 0.65487 regularization_factors, val_score: 0.654870: 20%|## | 4/20 [00:06<00:27, 1.71s/it][I 2020-09-27 04:46:16,718] Trial 46 finished with value: 0.654869889386949 and parameters: {'lambda_l1': 3.0122223559697e-08, 'lambda_l2': 0.00021523441226079608}. Best is trial 46 with value: 0.654869889386949. regularization_factors, val_score: 0.654870: 20%|## | 4/20 [00:07<00:27, 1.71s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000873 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.263726 valid's binary_logloss: 0.666214 Early stopping, best iteration is: [37] train's binary_logloss: 0.445141 valid's binary_logloss: 0.654998 regularization_factors, val_score: 0.654870: 25%|##5 | 5/20 [00:09<00:27, 1.80s/it][I 2020-09-27 04:46:18,736] Trial 47 finished with value: 0.6549975103511791 and parameters: {'lambda_l1': 5.742885425790403e-08, 'lambda_l2': 0.0002767008648395505}. Best is trial 46 with value: 0.654869889386949. regularization_factors, val_score: 0.654870: 25%|##5 | 5/20 [00:09<00:27, 1.80s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001046 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.264666 valid's binary_logloss: 0.666984 Early stopping, best iteration is: [37] train's binary_logloss: 0.445139 valid's binary_logloss: 0.654961 regularization_factors, val_score: 0.654870: 30%|### | 6/20 [00:10<00:24, 1.73s/it][I 2020-09-27 04:46:20,296] Trial 48 finished with value: 0.6549605608644221 and parameters: {'lambda_l1': 8.607940177177832e-08, 'lambda_l2': 0.0001457048379141767}. Best is trial 46 with value: 0.654869889386949. regularization_factors, val_score: 0.654870: 30%|### | 6/20 [00:10<00:24, 1.73s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000921 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.263723 valid's binary_logloss: 0.666273 Early stopping, best iteration is: [37] train's binary_logloss: 0.445139 valid's binary_logloss: 0.655013 regularization_factors, val_score: 0.654870: 35%|###5 | 7/20 [00:12<00:24, 1.87s/it][I 2020-09-27 04:46:22,480] Trial 49 finished with value: 0.6550133681715706 and parameters: {'lambda_l1': 1.014605269031525e-08, 'lambda_l2': 0.00017168614663450724}. Best is trial 46 with value: 0.654869889386949. regularization_factors, val_score: 0.654870: 35%|###5 | 7/20 [00:12<00:24, 1.87s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004784 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.263723 valid's binary_logloss: 0.666224 Early stopping, best iteration is: [37] train's binary_logloss: 0.445139 valid's binary_logloss: 0.654928 regularization_factors, val_score: 0.654870: 40%|#### | 8/20 [00:14<00:20, 1.72s/it][I 2020-09-27 04:46:23,853] Trial 50 finished with value: 0.6549280045293944 and parameters: {'lambda_l1': 1.4571400606028584e-08, 'lambda_l2': 0.00018545051962322963}. Best is trial 46 with value: 0.654869889386949. regularization_factors, val_score: 0.654870: 40%|#### | 8/20 [00:14<00:20, 1.72s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001114 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.263723 valid's binary_logloss: 0.666238 Early stopping, best iteration is: [37] train's binary_logloss: 0.445139 valid's binary_logloss: 0.654958 regularization_factors, val_score: 0.654870: 45%|####5 | 9/20 [00:16<00:20, 1.83s/it][I 2020-09-27 04:46:25,937] Trial 51 finished with value: 0.6549580786078956 and parameters: {'lambda_l1': 1.1439375782625129e-08, 'lambda_l2': 0.00020040952933499966}. Best is trial 46 with value: 0.654869889386949. regularization_factors, val_score: 0.654870: 45%|####5 | 9/20 [00:16<00:20, 1.83s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004784 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.264844 valid's binary_logloss: 0.665542 Early stopping, best iteration is: [43] train's binary_logloss: 0.421453 valid's binary_logloss: 0.654899 regularization_factors, val_score: 0.654870: 50%|##### | 10/20 [00:17<00:17, 1.71s/it][I 2020-09-27 04:46:27,383] Trial 52 finished with value: 0.6548985889877852 and parameters: {'lambda_l1': 1.3842789551680254e-08, 'lambda_l2': 5.6495755460565416e-05}. Best is trial 46 with value: 0.654869889386949. regularization_factors, val_score: 0.654870: 50%|##### | 10/20 [00:17<00:17, 1.71s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004465 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.264846 valid's binary_logloss: 0.665605 Early stopping, best iteration is: [38] train's binary_logloss: 0.441545 valid's binary_logloss: 0.654867 regularization_factors, val_score: 0.654867: 55%|#####5 | 11/20 [00:19<00:16, 1.80s/it][I 2020-09-27 04:46:29,389] Trial 53 finished with value: 0.6548668432849638 and parameters: {'lambda_l1': 1.189642528157414e-08, 'lambda_l2': 5.203275530782603e-06}. Best is trial 53 with value: 0.6548668432849638. regularization_factors, val_score: 0.654867: 55%|#####5 | 11/20 [00:19<00:16, 1.80s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000851 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.264846 valid's binary_logloss: 0.665516 Early stopping, best iteration is: [38] train's binary_logloss: 0.441545 valid's binary_logloss: 0.654838 regularization_factors, val_score: 0.654838: 60%|###### | 12/20 [00:21<00:13, 1.73s/it][I 2020-09-27 04:46:30,946] Trial 54 finished with value: 0.6548376800578837 and parameters: {'lambda_l1': 2.1670289548196517e-06, 'lambda_l2': 1.661838347245463e-06}. Best is trial 54 with value: 0.6548376800578837. regularization_factors, val_score: 0.654838: 60%|###### | 12/20 [00:21<00:13, 1.73s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000858 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.264846 valid's binary_logloss: 0.66563 Early stopping, best iteration is: [38] train's binary_logloss: 0.441545 valid's binary_logloss: 0.654905 regularization_factors, val_score: 0.654838: 65%|######5 | 13/20 [00:23<00:12, 1.84s/it][I 2020-09-27 04:46:33,033] Trial 55 finished with value: 0.6549047704174789 and parameters: {'lambda_l1': 1.3047485080865155e-05, 'lambda_l2': 1.1388372443834088e-06}. Best is trial 54 with value: 0.6548376800578837. regularization_factors, val_score: 0.654838: 65%|######5 | 13/20 [00:23<00:12, 1.84s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000859 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.264846 valid's binary_logloss: 0.665482 Early stopping, best iteration is: [38] train's binary_logloss: 0.441545 valid's binary_logloss: 0.654874 regularization_factors, val_score: 0.654838: 70%|####### | 14/20 [00:24<00:10, 1.75s/it][I 2020-09-27 04:46:34,599] Trial 56 finished with value: 0.6548744228748046 and parameters: {'lambda_l1': 1.677207413460934e-06, 'lambda_l2': 2.3547411212794384e-06}. Best is trial 54 with value: 0.6548376800578837. regularization_factors, val_score: 0.654838: 70%|####### | 14/20 [00:24<00:10, 1.75s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000869 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.264846 valid's binary_logloss: 0.665557 Early stopping, best iteration is: [38] train's binary_logloss: 0.441545 valid's binary_logloss: 0.654891 regularization_factors, val_score: 0.654838: 75%|#######5 | 15/20 [00:27<00:09, 1.87s/it][I 2020-09-27 04:46:36,725] Trial 57 finished with value: 0.65489106531928 and parameters: {'lambda_l1': 2.808990582697963e-06, 'lambda_l2': 3.680142856828128e-06}. Best is trial 54 with value: 0.6548376800578837. regularization_factors, val_score: 0.654838: 75%|#######5 | 15/20 [00:27<00:09, 1.87s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005001 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.264846 valid's binary_logloss: 0.665537 Early stopping, best iteration is: [38] train's binary_logloss: 0.441545 valid's binary_logloss: 0.654883 regularization_factors, val_score: 0.654838: 80%|######## | 16/20 [00:28<00:07, 1.75s/it][I 2020-09-27 04:46:38,214] Trial 58 finished with value: 0.6548833147276146 and parameters: {'lambda_l1': 3.548471771976571e-06, 'lambda_l2': 1.3069611448432158e-06}. Best is trial 54 with value: 0.6548376800578837. regularization_factors, val_score: 0.654838: 80%|######## | 16/20 [00:28<00:07, 1.75s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000875 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.264846 valid's binary_logloss: 0.665518 Early stopping, best iteration is: [38] train's binary_logloss: 0.441545 valid's binary_logloss: 0.654891 regularization_factors, val_score: 0.654838: 85%|########5 | 17/20 [00:30<00:05, 1.86s/it][I 2020-09-27 04:46:40,318] Trial 59 finished with value: 0.654890562836266 and parameters: {'lambda_l1': 3.6573101362801625e-06, 'lambda_l2': 1.3655499334616953e-06}. Best is trial 54 with value: 0.6548376800578837. regularization_factors, val_score: 0.654838: 85%|########5 | 17/20 [00:30<00:05, 1.86s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000836 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.263781 valid's binary_logloss: 0.676626 Early stopping, best iteration is: [41] train's binary_logloss: 0.42877 valid's binary_logloss: 0.654741 regularization_factors, val_score: 0.654741: 90%|######### | 18/20 [00:32<00:03, 1.78s/it][I 2020-09-27 04:46:41,916] Trial 60 finished with value: 0.6547406780914222 and parameters: {'lambda_l1': 8.237457561826076e-05, 'lambda_l2': 7.033731015507948e-07}. Best is trial 60 with value: 0.6547406780914222. regularization_factors, val_score: 0.654741: 90%|######### | 18/20 [00:32<00:03, 1.78s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000852 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.264845 valid's binary_logloss: 0.66557 Early stopping, best iteration is: [38] train's binary_logloss: 0.441546 valid's binary_logloss: 0.654905 regularization_factors, val_score: 0.654741: 95%|#########5| 19/20 [00:34<00:01, 1.87s/it][I 2020-09-27 04:46:43,993] Trial 61 finished with value: 0.6549054519516266 and parameters: {'lambda_l1': 8.016334404835216e-05, 'lambda_l2': 8.515948497846046e-07}. Best is trial 60 with value: 0.6547406780914222. regularization_factors, val_score: 0.654741: 95%|#########5| 19/20 [00:34<00:01, 1.87s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000852 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.264846 valid's binary_logloss: 0.665533 Early stopping, best iteration is: [38] train's binary_logloss: 0.441545 valid's binary_logloss: 0.65488 regularization_factors, val_score: 0.654741: 100%|##########| 20/20 [00:35<00:00, 1.78s/it][I 2020-09-27 04:46:45,569] Trial 62 finished with value: 0.6548801766746073 and parameters: {'lambda_l1': 6.807709893234408e-07, 'lambda_l2': 1.3368946882298463e-06}. Best is trial 60 with value: 0.6547406780914222. regularization_factors, val_score: 0.654741: 100%|##########| 20/20 [00:35<00:00, 1.79s/it] min_data_in_leaf, val_score: 0.654741: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000788 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.26674 valid's binary_logloss: 0.673371 Early stopping, best iteration is: [37] train's binary_logloss: 0.44854 valid's binary_logloss: 0.659997 min_data_in_leaf, val_score: 0.654741: 20%|## | 1/5 [00:02<00:08, 2.00s/it][I 2020-09-27 04:46:47,591] Trial 63 finished with value: 0.6599970673875724 and parameters: {'min_child_samples': 25}. Best is trial 63 with value: 0.6599970673875724. min_data_in_leaf, val_score: 0.654741: 20%|## | 1/5 [00:02<00:08, 2.00s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000851 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.259932 valid's binary_logloss: 0.674681 Early stopping, best iteration is: [27] train's binary_logloss: 0.485136 valid's binary_logloss: 0.663118 min_data_in_leaf, val_score: 0.654741: 40%|#### | 2/5 [00:03<00:05, 1.90s/it][I 2020-09-27 04:46:49,234] Trial 64 finished with value: 0.6631181004712279 and parameters: {'min_child_samples': 5}. Best is trial 63 with value: 0.6599970673875724. min_data_in_leaf, val_score: 0.654741: 40%|#### | 2/5 [00:03<00:05, 1.90s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000813 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.277711 valid's binary_logloss: 0.677648 Early stopping, best iteration is: [35] train's binary_logloss: 0.469589 valid's binary_logloss: 0.66334 min_data_in_leaf, val_score: 0.654741: 60%|###### | 3/5 [00:05<00:03, 1.93s/it][I 2020-09-27 04:46:51,228] Trial 65 finished with value: 0.6633400797460791 and parameters: {'min_child_samples': 50}. Best is trial 63 with value: 0.6599970673875724. min_data_in_leaf, val_score: 0.654741: 60%|###### | 3/5 [00:05<00:03, 1.93s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000892 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.262769 valid's binary_logloss: 0.674802 Early stopping, best iteration is: [35] train's binary_logloss: 0.44869 valid's binary_logloss: 0.662608 min_data_in_leaf, val_score: 0.654741: 80%|######## | 4/5 [00:07<00:01, 1.86s/it][I 2020-09-27 04:46:52,928] Trial 66 finished with value: 0.6626078522269558 and parameters: {'min_child_samples': 10}. Best is trial 63 with value: 0.6599970673875724. min_data_in_leaf, val_score: 0.654741: 80%|######## | 4/5 [00:07<00:01, 1.86s/it][LightGBM] [Info] Number of positive: 12849, number of negative: 13150 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000877 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4238 [LightGBM] [Info] Number of data points in the train set: 25999, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494211 -> initscore=-0.023156 [LightGBM] [Info] Start training from score -0.023156 [LightGBM] [Warning] No further splits with positive gain, best gain: -inf Training until validation scores don't improve for 100 rounds [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [100] train's binary_logloss: 0.355756 valid's binary_logloss: 0.678018 [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf [LightGBM] [Warning] No further splits with positive gain, best gain: -inf Early stopping, best iteration is: [32] train's binary_logloss: 0.530938 valid's binary_logloss: 0.659107 min_data_in_leaf, val_score: 0.654741: 100%|##########| 5/5 [00:10<00:00, 2.23s/it][I 2020-09-27 04:46:56,042] Trial 67 finished with value: 0.6591070735201429 and parameters: {'min_child_samples': 100}. Best is trial 67 with value: 0.6591070735201429. min_data_in_leaf, val_score: 0.654741: 100%|##########| 5/5 [00:10<00:00, 2.09s/it] Fold : 8 [I 2020-09-27 04:46:56,086] A new study created in memory with name: no-name-0ac6def8-9fc0-44ca-a007-4dfc936cce3b feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000362 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.582362 valid's binary_logloss: 0.657172 Early stopping, best iteration is: [75] train's binary_logloss: 0.598686 valid's binary_logloss: 0.656832 feature_fraction, val_score: 0.656832: 14%|#4 | 1/7 [00:00<00:03, 1.84it/s][I 2020-09-27 04:46:56,647] Trial 0 finished with value: 0.6568322981471987 and parameters: {'feature_fraction': 0.4}. Best is trial 0 with value: 0.6568322981471987. feature_fraction, val_score: 0.656832: 14%|#4 | 1/7 [00:00<00:03, 1.84it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004661 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.575682 valid's binary_logloss: 0.657591 Early stopping, best iteration is: [65] train's binary_logloss: 0.598895 valid's binary_logloss: 0.65488 feature_fraction, val_score: 0.654880: 29%|##8 | 2/7 [00:01<00:02, 1.90it/s][I 2020-09-27 04:46:57,132] Trial 1 finished with value: 0.654879918185127 and parameters: {'feature_fraction': 0.7}. Best is trial 1 with value: 0.654879918185127. feature_fraction, val_score: 0.654880: 29%|##8 | 2/7 [00:01<00:02, 1.90it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000886 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.573899 valid's binary_logloss: 0.657206 Early stopping, best iteration is: [70] train's binary_logloss: 0.593743 valid's binary_logloss: 0.655846 feature_fraction, val_score: 0.654880: 43%|####2 | 3/7 [00:01<00:02, 1.85it/s][I 2020-09-27 04:46:57,704] Trial 2 finished with value: 0.6558455623356558 and parameters: {'feature_fraction': 0.8}. Best is trial 1 with value: 0.654879918185127. feature_fraction, val_score: 0.654880: 43%|####2 | 3/7 [00:01<00:02, 1.85it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000972 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.57084 valid's binary_logloss: 0.654546 [200] train's binary_logloss: 0.513329 valid's binary_logloss: 0.658828 Early stopping, best iteration is: [113] train's binary_logloss: 0.562928 valid's binary_logloss: 0.654073 feature_fraction, val_score: 0.654073: 57%|#####7 | 4/7 [00:02<00:01, 1.73it/s][I 2020-09-27 04:46:58,369] Trial 3 finished with value: 0.6540730143979465 and parameters: {'feature_fraction': 1.0}. Best is trial 3 with value: 0.6540730143979465. feature_fraction, val_score: 0.654073: 57%|#####7 | 4/7 [00:02<00:01, 1.73it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004601 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.579014 valid's binary_logloss: 0.655596 Early stopping, best iteration is: [59] train's binary_logloss: 0.606682 valid's binary_logloss: 0.6549 feature_fraction, val_score: 0.654073: 71%|#######1 | 5/7 [00:02<00:01, 1.63it/s][I 2020-09-27 04:46:59,068] Trial 4 finished with value: 0.6548998752828071 and parameters: {'feature_fraction': 0.5}. Best is trial 3 with value: 0.6540730143979465. feature_fraction, val_score: 0.654073: 71%|#######1 | 5/7 [00:02<00:01, 1.63it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011929 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.572184 valid's binary_logloss: 0.658179 Early stopping, best iteration is: [67] train's binary_logloss: 0.594979 valid's binary_logloss: 0.656087 feature_fraction, val_score: 0.654073: 86%|########5 | 6/7 [00:03<00:00, 1.52it/s][I 2020-09-27 04:46:59,834] Trial 5 finished with value: 0.6560873120955287 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 3 with value: 0.6540730143979465. feature_fraction, val_score: 0.654073: 86%|########5 | 6/7 [00:03<00:00, 1.52it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004851 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.576475 valid's binary_logloss: 0.652717 Early stopping, best iteration is: [82] train's binary_logloss: 0.587886 valid's binary_logloss: 0.65223 feature_fraction, val_score: 0.652230: 100%|##########| 7/7 [00:04<00:00, 1.64it/s][I 2020-09-27 04:47:00,328] Trial 6 finished with value: 0.6522296496116577 and parameters: {'feature_fraction': 0.6}. Best is trial 6 with value: 0.6522296496116577. feature_fraction, val_score: 0.652230: 100%|##########| 7/7 [00:04<00:00, 1.65it/s] num_leaves, val_score: 0.652230: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004842 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.326219 valid's binary_logloss: 0.670645 Early stopping, best iteration is: [35] train's binary_logloss: 0.497287 valid's binary_logloss: 0.658257 num_leaves, val_score: 0.652230: 5%|5 | 1/20 [00:00<00:17, 1.08it/s][I 2020-09-27 04:47:01,277] Trial 7 finished with value: 0.6582573078988242 and parameters: {'num_leaves': 200}. Best is trial 7 with value: 0.6582573078988242. num_leaves, val_score: 0.652230: 5%|5 | 1/20 [00:00<00:17, 1.08it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004568 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.281259 valid's binary_logloss: 0.679658 Early stopping, best iteration is: [36] train's binary_logloss: 0.462423 valid's binary_logloss: 0.663543 num_leaves, val_score: 0.652230: 10%|# | 2/20 [00:02<00:17, 1.03it/s][I 2020-09-27 04:47:02,352] Trial 8 finished with value: 0.6635430249480666 and parameters: {'num_leaves': 251}. Best is trial 7 with value: 0.6582573078988242. num_leaves, val_score: 0.652230: 10%|# | 2/20 [00:02<00:17, 1.03it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005025 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657954 valid's binary_logloss: 0.664118 [200] train's binary_logloss: 0.647761 valid's binary_logloss: 0.655763 [300] train's binary_logloss: 0.642622 valid's binary_logloss: 0.652688 [400] train's binary_logloss: 0.638865 valid's binary_logloss: 0.651406 [500] train's binary_logloss: 0.635649 valid's binary_logloss: 0.650648 [600] train's binary_logloss: 0.632824 valid's binary_logloss: 0.650562 [700] train's binary_logloss: 0.630241 valid's binary_logloss: 0.650393 Early stopping, best iteration is: [694] train's binary_logloss: 0.630392 valid's binary_logloss: 0.650237 num_leaves, val_score: 0.650237: 15%|#5 | 3/20 [00:03<00:19, 1.17s/it][I 2020-09-27 04:47:03,969] Trial 9 finished with value: 0.65023693049223 and parameters: {'num_leaves': 3}. Best is trial 9 with value: 0.65023693049223. num_leaves, val_score: 0.650237: 15%|#5 | 3/20 [00:03<00:19, 1.17s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004669 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657954 valid's binary_logloss: 0.664118 [200] train's binary_logloss: 0.647761 valid's binary_logloss: 0.655763 [300] train's binary_logloss: 0.642622 valid's binary_logloss: 0.652688 [400] train's binary_logloss: 0.638865 valid's binary_logloss: 0.651406 [500] train's binary_logloss: 0.635649 valid's binary_logloss: 0.650648 [600] train's binary_logloss: 0.632824 valid's binary_logloss: 0.650562 [700] train's binary_logloss: 0.630241 valid's binary_logloss: 0.650393 Early stopping, best iteration is: [694] train's binary_logloss: 0.630392 valid's binary_logloss: 0.650237 num_leaves, val_score: 0.650237: 20%|## | 4/20 [00:04<00:18, 1.18s/it][I 2020-09-27 04:47:05,175] Trial 10 finished with value: 0.6502369304922299 and parameters: {'num_leaves': 3}. Best is trial 10 with value: 0.6502369304922299. num_leaves, val_score: 0.650237: 20%|## | 4/20 [00:04<00:18, 1.18s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005031 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.628976 valid's binary_logloss: 0.654294 [200] train's binary_logloss: 0.608343 valid's binary_logloss: 0.65336 Early stopping, best iteration is: [176] train's binary_logloss: 0.613043 valid's binary_logloss: 0.652178 num_leaves, val_score: 0.650237: 25%|##5 | 5/20 [00:05<00:14, 1.01it/s][I 2020-09-27 04:47:05,732] Trial 11 finished with value: 0.6521775910881051 and parameters: {'num_leaves': 10}. Best is trial 10 with value: 0.6502369304922299. num_leaves, val_score: 0.650237: 25%|##5 | 5/20 [00:05<00:14, 1.01it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004975 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.635408 valid's binary_logloss: 0.65585 [200] train's binary_logloss: 0.618644 valid's binary_logloss: 0.654948 Early stopping, best iteration is: [161] train's binary_logloss: 0.624573 valid's binary_logloss: 0.6542 num_leaves, val_score: 0.650237: 30%|### | 6/20 [00:05<00:11, 1.17it/s][I 2020-09-27 04:47:06,273] Trial 12 finished with value: 0.6541997386028066 and parameters: {'num_leaves': 8}. Best is trial 10 with value: 0.6502369304922299. num_leaves, val_score: 0.650237: 30%|### | 6/20 [00:05<00:11, 1.17it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000898 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.488946 valid's binary_logloss: 0.659659 Early stopping, best iteration is: [39] train's binary_logloss: 0.5773 valid's binary_logloss: 0.65642 num_leaves, val_score: 0.650237: 35%|###5 | 7/20 [00:07<00:12, 1.04it/s][I 2020-09-27 04:47:07,468] Trial 13 finished with value: 0.6564196118293095 and parameters: {'num_leaves': 76}. Best is trial 10 with value: 0.6502369304922299. num_leaves, val_score: 0.650237: 35%|###5 | 7/20 [00:07<00:12, 1.04it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000884 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.498154 valid's binary_logloss: 0.662208 Early stopping, best iteration is: [39] train's binary_logloss: 0.581162 valid's binary_logloss: 0.657338 num_leaves, val_score: 0.650237: 40%|#### | 8/20 [00:07<00:10, 1.16it/s][I 2020-09-27 04:47:08,100] Trial 14 finished with value: 0.6573383283023841 and parameters: {'num_leaves': 71}. Best is trial 10 with value: 0.6502369304922299. num_leaves, val_score: 0.650237: 40%|#### | 8/20 [00:07<00:10, 1.16it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004661 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.646593 valid's binary_logloss: 0.657131 [200] train's binary_logloss: 0.634403 valid's binary_logloss: 0.652331 [300] train's binary_logloss: 0.625932 valid's binary_logloss: 0.651212 Early stopping, best iteration is: [289] train's binary_logloss: 0.626825 valid's binary_logloss: 0.650977 num_leaves, val_score: 0.650237: 45%|####5 | 9/20 [00:08<00:08, 1.27it/s][I 2020-09-27 04:47:08,723] Trial 15 finished with value: 0.6509770662129648 and parameters: {'num_leaves': 5}. Best is trial 10 with value: 0.6502369304922299. num_leaves, val_score: 0.650237: 45%|####5 | 9/20 [00:08<00:08, 1.27it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004588 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.498154 valid's binary_logloss: 0.662208 Early stopping, best iteration is: [39] train's binary_logloss: 0.581162 valid's binary_logloss: 0.657338 num_leaves, val_score: 0.650237: 50%|##### | 10/20 [00:08<00:07, 1.40it/s][I 2020-09-27 04:47:09,256] Trial 16 finished with value: 0.6573383283023841 and parameters: {'num_leaves': 71}. Best is trial 10 with value: 0.6502369304922299. num_leaves, val_score: 0.650237: 50%|##### | 10/20 [00:08<00:07, 1.40it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012770 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.392369 valid's binary_logloss: 0.669429 Early stopping, best iteration is: [45] train's binary_logloss: 0.507218 valid's binary_logloss: 0.65932 num_leaves, val_score: 0.650237: 55%|#####5 | 11/20 [00:09<00:06, 1.35it/s][I 2020-09-27 04:47:10,058] Trial 17 finished with value: 0.659319879723429 and parameters: {'num_leaves': 141}. Best is trial 10 with value: 0.6502369304922299. num_leaves, val_score: 0.650237: 55%|#####5 | 11/20 [00:09<00:06, 1.35it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010521 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.553596 valid's binary_logloss: 0.658319 Early stopping, best iteration is: [54] train's binary_logloss: 0.593584 valid's binary_logloss: 0.656096 num_leaves, val_score: 0.650237: 60%|###### | 12/20 [00:10<00:06, 1.25it/s][I 2020-09-27 04:47:10,992] Trial 18 finished with value: 0.6560958307756177 and parameters: {'num_leaves': 42}. Best is trial 10 with value: 0.6502369304922299. num_leaves, val_score: 0.650237: 60%|###### | 12/20 [00:10<00:06, 1.25it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001062 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.409616 valid's binary_logloss: 0.665328 Early stopping, best iteration is: [37] train's binary_logloss: 0.53918 valid's binary_logloss: 0.658274 num_leaves, val_score: 0.650237: 65%|######5 | 13/20 [00:11<00:05, 1.19it/s][I 2020-09-27 04:47:11,928] Trial 19 finished with value: 0.6582743650249848 and parameters: {'num_leaves': 128}. Best is trial 10 with value: 0.6502369304922299. num_leaves, val_score: 0.650237: 65%|######5 | 13/20 [00:11<00:05, 1.19it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004694 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.563913 valid's binary_logloss: 0.657197 Early stopping, best iteration is: [77] train's binary_logloss: 0.580794 valid's binary_logloss: 0.654988 num_leaves, val_score: 0.650237: 70%|####### | 14/20 [00:12<00:04, 1.37it/s][I 2020-09-27 04:47:12,412] Trial 20 finished with value: 0.6549884052454669 and parameters: {'num_leaves': 37}. Best is trial 10 with value: 0.6502369304922299. num_leaves, val_score: 0.650237: 70%|####### | 14/20 [00:12<00:04, 1.37it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009279 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657954 valid's binary_logloss: 0.664118 [200] train's binary_logloss: 0.647761 valid's binary_logloss: 0.655763 [300] train's binary_logloss: 0.642622 valid's binary_logloss: 0.652688 [400] train's binary_logloss: 0.638865 valid's binary_logloss: 0.651406 [500] train's binary_logloss: 0.635649 valid's binary_logloss: 0.650648 [600] train's binary_logloss: 0.632824 valid's binary_logloss: 0.650562 [700] train's binary_logloss: 0.630241 valid's binary_logloss: 0.650393 Early stopping, best iteration is: [694] train's binary_logloss: 0.630392 valid's binary_logloss: 0.650237 num_leaves, val_score: 0.650237: 75%|#######5 | 15/20 [00:13<00:04, 1.18it/s][I 2020-09-27 04:47:13,534] Trial 21 finished with value: 0.6502369304922299 and parameters: {'num_leaves': 3}. Best is trial 10 with value: 0.6502369304922299. num_leaves, val_score: 0.650237: 75%|#######5 | 15/20 [00:13<00:04, 1.18it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000785 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.646593 valid's binary_logloss: 0.657131 [200] train's binary_logloss: 0.634403 valid's binary_logloss: 0.652331 [300] train's binary_logloss: 0.625932 valid's binary_logloss: 0.651212 Early stopping, best iteration is: [289] train's binary_logloss: 0.626825 valid's binary_logloss: 0.650977 num_leaves, val_score: 0.650237: 80%|######## | 16/20 [00:14<00:03, 1.08it/s][I 2020-09-27 04:47:14,641] Trial 22 finished with value: 0.6509770662129648 and parameters: {'num_leaves': 5}. Best is trial 10 with value: 0.6502369304922299. num_leaves, val_score: 0.650237: 80%|######## | 16/20 [00:14<00:03, 1.08it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.019722 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.565177 valid's binary_logloss: 0.655115 Early stopping, best iteration is: [72] train's binary_logloss: 0.585373 valid's binary_logloss: 0.65339 num_leaves, val_score: 0.650237: 85%|########5 | 17/20 [00:14<00:02, 1.20it/s][I 2020-09-27 04:47:15,254] Trial 23 finished with value: 0.6533900763638031 and parameters: {'num_leaves': 36}. Best is trial 10 with value: 0.6502369304922299. num_leaves, val_score: 0.650237: 85%|########5 | 17/20 [00:14<00:02, 1.20it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004882 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.565177 valid's binary_logloss: 0.655115 Early stopping, best iteration is: [72] train's binary_logloss: 0.585373 valid's binary_logloss: 0.65339 num_leaves, val_score: 0.650237: 90%|######### | 18/20 [00:15<00:01, 1.36it/s][I 2020-09-27 04:47:15,759] Trial 24 finished with value: 0.6533900763638031 and parameters: {'num_leaves': 36}. Best is trial 10 with value: 0.6502369304922299. num_leaves, val_score: 0.650237: 90%|######### | 18/20 [00:15<00:01, 1.36it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009941 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.668764 valid's binary_logloss: 0.672745 [200] train's binary_logloss: 0.65923 valid's binary_logloss: 0.663108 [300] train's binary_logloss: 0.654067 valid's binary_logloss: 0.658349 [400] train's binary_logloss: 0.651004 valid's binary_logloss: 0.655269 [500] train's binary_logloss: 0.649032 valid's binary_logloss: 0.653548 [600] train's binary_logloss: 0.647704 valid's binary_logloss: 0.652575 [700] train's binary_logloss: 0.646761 valid's binary_logloss: 0.651831 [800] train's binary_logloss: 0.646051 valid's binary_logloss: 0.651481 [900] train's binary_logloss: 0.645487 valid's binary_logloss: 0.651115 [1000] train's binary_logloss: 0.645019 valid's binary_logloss: 0.651086 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.645019 valid's binary_logloss: 0.651086 num_leaves, val_score: 0.650237: 95%|#########5| 19/20 [00:16<00:00, 1.12it/s][I 2020-09-27 04:47:17,024] Trial 25 finished with value: 0.6510861135332809 and parameters: {'num_leaves': 2}. Best is trial 10 with value: 0.6502369304922299. num_leaves, val_score: 0.650237: 95%|#########5| 19/20 [00:16<00:00, 1.12it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005016 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.424526 valid's binary_logloss: 0.659895 Early stopping, best iteration is: [33] train's binary_logloss: 0.556295 valid's binary_logloss: 0.654704 num_leaves, val_score: 0.650237: 100%|##########| 20/20 [00:17<00:00, 1.20it/s][I 2020-09-27 04:47:17,712] Trial 26 finished with value: 0.6547044966874118 and parameters: {'num_leaves': 119}. Best is trial 10 with value: 0.6502369304922299. num_leaves, val_score: 0.650237: 100%|##########| 20/20 [00:17<00:00, 1.15it/s] bagging, val_score: 0.650237: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004738 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.658045 valid's binary_logloss: 0.664127 [200] train's binary_logloss: 0.647774 valid's binary_logloss: 0.655872 [300] train's binary_logloss: 0.642497 valid's binary_logloss: 0.65234 [400] train's binary_logloss: 0.638669 valid's binary_logloss: 0.651347 [500] train's binary_logloss: 0.635478 valid's binary_logloss: 0.651168 Early stopping, best iteration is: [453] train's binary_logloss: 0.63696 valid's binary_logloss: 0.650827 bagging, val_score: 0.650237: 10%|# | 1/10 [00:01<00:12, 1.44s/it][I 2020-09-27 04:47:19,168] Trial 27 finished with value: 0.6508272913841253 and parameters: {'bagging_fraction': 0.9963833291336945, 'bagging_freq': 6}. Best is trial 27 with value: 0.6508272913841253. bagging, val_score: 0.650237: 10%|# | 1/10 [00:01<00:12, 1.44s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004596 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.655889 valid's binary_logloss: 0.659928 [200] train's binary_logloss: 0.646169 valid's binary_logloss: 0.654108 [300] train's binary_logloss: 0.641381 valid's binary_logloss: 0.651946 [400] train's binary_logloss: 0.637836 valid's binary_logloss: 0.650942 Early stopping, best iteration is: [389] train's binary_logloss: 0.638252 valid's binary_logloss: 0.650231 bagging, val_score: 0.650231: 20%|## | 2/10 [00:02<00:10, 1.27s/it][I 2020-09-27 04:47:20,036] Trial 28 finished with value: 0.6502306281879949 and parameters: {'bagging_fraction': 0.4514619063105707, 'bagging_freq': 1}. Best is trial 28 with value: 0.6502306281879949. bagging, val_score: 0.650231: 20%|## | 2/10 [00:02<00:10, 1.27s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004540 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.655825 valid's binary_logloss: 0.660152 [200] train's binary_logloss: 0.646253 valid's binary_logloss: 0.653794 [300] train's binary_logloss: 0.64168 valid's binary_logloss: 0.652426 [400] train's binary_logloss: 0.638103 valid's binary_logloss: 0.651436 Early stopping, best iteration is: [381] train's binary_logloss: 0.63871 valid's binary_logloss: 0.650844 bagging, val_score: 0.650231: 30%|### | 3/10 [00:03<00:07, 1.13s/it][I 2020-09-27 04:47:20,849] Trial 29 finished with value: 0.6508442537589473 and parameters: {'bagging_fraction': 0.4102073189555854, 'bagging_freq': 1}. Best is trial 28 with value: 0.6502306281879949. bagging, val_score: 0.650231: 30%|### | 3/10 [00:03<00:07, 1.13s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004944 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.655879 valid's binary_logloss: 0.660227 [200] train's binary_logloss: 0.646389 valid's binary_logloss: 0.653638 [300] train's binary_logloss: 0.641982 valid's binary_logloss: 0.652522 [400] train's binary_logloss: 0.638289 valid's binary_logloss: 0.652467 Early stopping, best iteration is: [387] train's binary_logloss: 0.63872 valid's binary_logloss: 0.652011 bagging, val_score: 0.650231: 40%|#### | 4/10 [00:03<00:06, 1.04s/it][I 2020-09-27 04:47:21,683] Trial 30 finished with value: 0.6520113408937251 and parameters: {'bagging_fraction': 0.4058702957626291, 'bagging_freq': 1}. Best is trial 28 with value: 0.6502306281879949. bagging, val_score: 0.650231: 40%|#### | 4/10 [00:03<00:06, 1.04s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008559 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656756 valid's binary_logloss: 0.662314 [200] train's binary_logloss: 0.646612 valid's binary_logloss: 0.653269 [300] train's binary_logloss: 0.641725 valid's binary_logloss: 0.650745 [400] train's binary_logloss: 0.638127 valid's binary_logloss: 0.650437 Early stopping, best iteration is: [386] train's binary_logloss: 0.638596 valid's binary_logloss: 0.649682 bagging, val_score: 0.649682: 50%|##### | 5/10 [00:05<00:05, 1.10s/it][I 2020-09-27 04:47:22,921] Trial 31 finished with value: 0.6496818435732342 and parameters: {'bagging_fraction': 0.6494590447691004, 'bagging_freq': 3}. Best is trial 31 with value: 0.6496818435732342. bagging, val_score: 0.649682: 50%|##### | 5/10 [00:05<00:05, 1.10s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004698 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.65681 valid's binary_logloss: 0.661741 [200] train's binary_logloss: 0.646579 valid's binary_logloss: 0.653526 [300] train's binary_logloss: 0.641414 valid's binary_logloss: 0.651164 [400] train's binary_logloss: 0.637939 valid's binary_logloss: 0.650273 Early stopping, best iteration is: [386] train's binary_logloss: 0.638417 valid's binary_logloss: 0.649953 bagging, val_score: 0.649682: 60%|###### | 6/10 [00:05<00:03, 1.00it/s][I 2020-09-27 04:47:23,670] Trial 32 finished with value: 0.6499526525656477 and parameters: {'bagging_fraction': 0.6912299232252367, 'bagging_freq': 3}. Best is trial 31 with value: 0.6496818435732342. bagging, val_score: 0.649682: 60%|###### | 6/10 [00:05<00:03, 1.00it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004577 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656917 valid's binary_logloss: 0.662005 [200] train's binary_logloss: 0.646694 valid's binary_logloss: 0.653638 [300] train's binary_logloss: 0.641556 valid's binary_logloss: 0.651449 [400] train's binary_logloss: 0.637865 valid's binary_logloss: 0.650759 Early stopping, best iteration is: [387] train's binary_logloss: 0.638305 valid's binary_logloss: 0.650278 bagging, val_score: 0.649682: 70%|####### | 7/10 [00:06<00:02, 1.09it/s][I 2020-09-27 04:47:24,413] Trial 33 finished with value: 0.6502781308682245 and parameters: {'bagging_fraction': 0.7006917452683191, 'bagging_freq': 3}. Best is trial 31 with value: 0.6496818435732342. bagging, val_score: 0.649682: 70%|####### | 7/10 [00:06<00:02, 1.09it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005647 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656936 valid's binary_logloss: 0.66276 [200] train's binary_logloss: 0.646675 valid's binary_logloss: 0.653779 [300] train's binary_logloss: 0.641779 valid's binary_logloss: 0.651157 [400] train's binary_logloss: 0.638073 valid's binary_logloss: 0.650615 Early stopping, best iteration is: [385] train's binary_logloss: 0.638614 valid's binary_logloss: 0.649988 bagging, val_score: 0.649682: 80%|######## | 8/10 [00:07<00:01, 1.14it/s][I 2020-09-27 04:47:25,180] Trial 34 finished with value: 0.6499876467524935 and parameters: {'bagging_fraction': 0.6697359412743568, 'bagging_freq': 3}. Best is trial 31 with value: 0.6496818435732342. bagging, val_score: 0.649682: 80%|######## | 8/10 [00:07<00:01, 1.14it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005015 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656928 valid's binary_logloss: 0.661654 [200] train's binary_logloss: 0.646722 valid's binary_logloss: 0.652579 [300] train's binary_logloss: 0.641759 valid's binary_logloss: 0.65045 [400] train's binary_logloss: 0.638172 valid's binary_logloss: 0.649763 [500] train's binary_logloss: 0.634789 valid's binary_logloss: 0.649596 Early stopping, best iteration is: [472] train's binary_logloss: 0.635712 valid's binary_logloss: 0.649063 bagging, val_score: 0.649063: 90%|######### | 9/10 [00:08<00:01, 1.00s/it][I 2020-09-27 04:47:26,487] Trial 35 finished with value: 0.649063213862809 and parameters: {'bagging_fraction': 0.6949061090676492, 'bagging_freq': 3}. Best is trial 35 with value: 0.649063213862809. bagging, val_score: 0.649063: 90%|######### | 9/10 [00:08<00:01, 1.00s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.006251 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656842 valid's binary_logloss: 0.662647 [200] train's binary_logloss: 0.646752 valid's binary_logloss: 0.654227 [300] train's binary_logloss: 0.641722 valid's binary_logloss: 0.65187 [400] train's binary_logloss: 0.638055 valid's binary_logloss: 0.651604 [500] train's binary_logloss: 0.634682 valid's binary_logloss: 0.651328 Early stopping, best iteration is: [464] train's binary_logloss: 0.63593 valid's binary_logloss: 0.650751 bagging, val_score: 0.649063: 100%|##########| 10/10 [00:09<00:00, 1.01s/it][I 2020-09-27 04:47:27,517] Trial 36 finished with value: 0.6507510423009876 and parameters: {'bagging_fraction': 0.6850010940586274, 'bagging_freq': 3}. Best is trial 35 with value: 0.649063213862809. bagging, val_score: 0.649063: 100%|##########| 10/10 [00:09<00:00, 1.02it/s] feature_fraction_stage2, val_score: 0.649063: 0%| | 0/6 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005016 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656928 valid's binary_logloss: 0.661654 [200] train's binary_logloss: 0.646722 valid's binary_logloss: 0.652579 [300] train's binary_logloss: 0.641759 valid's binary_logloss: 0.65045 [400] train's binary_logloss: 0.638172 valid's binary_logloss: 0.649763 [500] train's binary_logloss: 0.634789 valid's binary_logloss: 0.649596 Early stopping, best iteration is: [472] train's binary_logloss: 0.635712 valid's binary_logloss: 0.649063 feature_fraction_stage2, val_score: 0.649063: 17%|#6 | 1/6 [00:00<00:04, 1.11it/s][I 2020-09-27 04:47:28,434] Trial 37 finished with value: 0.649063213862809 and parameters: {'feature_fraction': 0.616}. Best is trial 37 with value: 0.649063213862809. feature_fraction_stage2, val_score: 0.649063: 17%|#6 | 1/6 [00:00<00:04, 1.11it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004551 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656692 valid's binary_logloss: 0.661077 [200] train's binary_logloss: 0.646429 valid's binary_logloss: 0.652636 [300] train's binary_logloss: 0.641474 valid's binary_logloss: 0.650303 [400] train's binary_logloss: 0.637696 valid's binary_logloss: 0.649448 [500] train's binary_logloss: 0.634319 valid's binary_logloss: 0.64952 Early stopping, best iteration is: [468] train's binary_logloss: 0.635429 valid's binary_logloss: 0.648813 feature_fraction_stage2, val_score: 0.648813: 33%|###3 | 2/6 [00:01<00:03, 1.11it/s][I 2020-09-27 04:47:29,319] Trial 38 finished with value: 0.6488132191521305 and parameters: {'feature_fraction': 0.6479999999999999}. Best is trial 38 with value: 0.6488132191521305. feature_fraction_stage2, val_score: 0.648813: 33%|###3 | 2/6 [00:01<00:03, 1.11it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000576 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657062 valid's binary_logloss: 0.661671 [200] train's binary_logloss: 0.647032 valid's binary_logloss: 0.653351 [300] train's binary_logloss: 0.641928 valid's binary_logloss: 0.650494 [400] train's binary_logloss: 0.638384 valid's binary_logloss: 0.649678 [500] train's binary_logloss: 0.634981 valid's binary_logloss: 0.649656 [600] train's binary_logloss: 0.631952 valid's binary_logloss: 0.649533 Early stopping, best iteration is: [551] train's binary_logloss: 0.633314 valid's binary_logloss: 0.649149 feature_fraction_stage2, val_score: 0.648813: 50%|##### | 3/6 [00:03<00:03, 1.06s/it][I 2020-09-27 04:47:30,759] Trial 39 finished with value: 0.6491494764152731 and parameters: {'feature_fraction': 0.552}. Best is trial 38 with value: 0.6488132191521305. feature_fraction_stage2, val_score: 0.648813: 50%|##### | 3/6 [00:03<00:03, 1.06s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000576 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656892 valid's binary_logloss: 0.661978 [200] train's binary_logloss: 0.646753 valid's binary_logloss: 0.653932 [300] train's binary_logloss: 0.641708 valid's binary_logloss: 0.652504 [400] train's binary_logloss: 0.638016 valid's binary_logloss: 0.651 [500] train's binary_logloss: 0.634749 valid's binary_logloss: 0.651944 Early stopping, best iteration is: [412] train's binary_logloss: 0.637606 valid's binary_logloss: 0.65077 feature_fraction_stage2, val_score: 0.648813: 67%|######6 | 4/6 [00:04<00:02, 1.00s/it][I 2020-09-27 04:47:31,621] Trial 40 finished with value: 0.6507703781531419 and parameters: {'feature_fraction': 0.584}. Best is trial 38 with value: 0.6488132191521305. feature_fraction_stage2, val_score: 0.648813: 67%|######6 | 4/6 [00:04<00:02, 1.00s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004974 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656759 valid's binary_logloss: 0.661816 [200] train's binary_logloss: 0.646462 valid's binary_logloss: 0.653161 [300] train's binary_logloss: 0.641483 valid's binary_logloss: 0.650694 [400] train's binary_logloss: 0.6377 valid's binary_logloss: 0.649329 [500] train's binary_logloss: 0.634322 valid's binary_logloss: 0.649442 Early stopping, best iteration is: [458] train's binary_logloss: 0.635675 valid's binary_logloss: 0.648669 feature_fraction_stage2, val_score: 0.648669: 83%|########3 | 5/6 [00:05<00:00, 1.01it/s][I 2020-09-27 04:47:32,586] Trial 41 finished with value: 0.6486693704765705 and parameters: {'feature_fraction': 0.6799999999999999}. Best is trial 41 with value: 0.6486693704765705. feature_fraction_stage2, val_score: 0.648669: 83%|########3 | 5/6 [00:05<00:00, 1.01it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004631 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657062 valid's binary_logloss: 0.661671 [200] train's binary_logloss: 0.647032 valid's binary_logloss: 0.653351 [300] train's binary_logloss: 0.641928 valid's binary_logloss: 0.650494 [400] train's binary_logloss: 0.638384 valid's binary_logloss: 0.649678 [500] train's binary_logloss: 0.634981 valid's binary_logloss: 0.649656 [600] train's binary_logloss: 0.631952 valid's binary_logloss: 0.649533 Early stopping, best iteration is: [551] train's binary_logloss: 0.633314 valid's binary_logloss: 0.649149 feature_fraction_stage2, val_score: 0.648669: 100%|##########| 6/6 [00:06<00:00, 1.11s/it][I 2020-09-27 04:47:33,975] Trial 42 finished with value: 0.6491494764152731 and parameters: {'feature_fraction': 0.52}. Best is trial 41 with value: 0.6486693704765705. feature_fraction_stage2, val_score: 0.648669: 100%|##########| 6/6 [00:06<00:00, 1.08s/it] regularization_factors, val_score: 0.648669: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012670 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656679 valid's binary_logloss: 0.661907 [200] train's binary_logloss: 0.646312 valid's binary_logloss: 0.653757 [300] train's binary_logloss: 0.641354 valid's binary_logloss: 0.651207 [400] train's binary_logloss: 0.637772 valid's binary_logloss: 0.650071 [500] train's binary_logloss: 0.634407 valid's binary_logloss: 0.649928 Early stopping, best iteration is: [473] train's binary_logloss: 0.635291 valid's binary_logloss: 0.649607 regularization_factors, val_score: 0.648669: 5%|5 | 1/20 [00:01<00:20, 1.08s/it][I 2020-09-27 04:47:35,081] Trial 43 finished with value: 0.6496074604263665 and parameters: {'lambda_l1': 4.193685786846261e-08, 'lambda_l2': 0.3229663189729637}. Best is trial 43 with value: 0.6496074604263665. regularization_factors, val_score: 0.648669: 5%|5 | 1/20 [00:01<00:20, 1.08s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011232 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657809 valid's binary_logloss: 0.663236 [200] train's binary_logloss: 0.647792 valid's binary_logloss: 0.654479 [300] train's binary_logloss: 0.643617 valid's binary_logloss: 0.650947 [400] train's binary_logloss: 0.640751 valid's binary_logloss: 0.650509 [500] train's binary_logloss: 0.638188 valid's binary_logloss: 0.650185 Early stopping, best iteration is: [464] train's binary_logloss: 0.639104 valid's binary_logloss: 0.649722 regularization_factors, val_score: 0.648669: 10%|# | 2/20 [00:01<00:18, 1.03s/it][I 2020-09-27 04:47:35,996] Trial 44 finished with value: 0.6497218154257693 and parameters: {'lambda_l1': 7.661272890098566, 'lambda_l2': 1.6647681314940023e-07}. Best is trial 43 with value: 0.6496074604263665. regularization_factors, val_score: 0.648669: 10%|# | 2/20 [00:02<00:18, 1.03s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005094 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656808 valid's binary_logloss: 0.661435 [200] train's binary_logloss: 0.646494 valid's binary_logloss: 0.653055 [300] train's binary_logloss: 0.641675 valid's binary_logloss: 0.650859 [400] train's binary_logloss: 0.637945 valid's binary_logloss: 0.649955 [500] train's binary_logloss: 0.634406 valid's binary_logloss: 0.649662 Early stopping, best iteration is: [469] train's binary_logloss: 0.635491 valid's binary_logloss: 0.649128 regularization_factors, val_score: 0.648669: 15%|#5 | 3/20 [00:02<00:17, 1.01s/it][I 2020-09-27 04:47:36,939] Trial 45 finished with value: 0.6491279609766097 and parameters: {'lambda_l1': 0.038151596318754866, 'lambda_l2': 1.3493847157523479e-08}. Best is trial 45 with value: 0.6491279609766097. regularization_factors, val_score: 0.648669: 15%|#5 | 3/20 [00:02<00:17, 1.01s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000809 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656763 valid's binary_logloss: 0.661526 [200] train's binary_logloss: 0.646359 valid's binary_logloss: 0.653842 [300] train's binary_logloss: 0.641324 valid's binary_logloss: 0.650914 [400] train's binary_logloss: 0.637727 valid's binary_logloss: 0.650544 [500] train's binary_logloss: 0.634443 valid's binary_logloss: 0.65063 Early stopping, best iteration is: [457] train's binary_logloss: 0.635863 valid's binary_logloss: 0.649897 regularization_factors, val_score: 0.648669: 20%|## | 4/20 [00:04<00:17, 1.12s/it][I 2020-09-27 04:47:38,329] Trial 46 finished with value: 0.6498965604608956 and parameters: {'lambda_l1': 0.061724342064673816, 'lambda_l2': 1.1191098300700474e-08}. Best is trial 45 with value: 0.6491279609766097. regularization_factors, val_score: 0.648669: 20%|## | 4/20 [00:04<00:17, 1.12s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005050 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656759 valid's binary_logloss: 0.661816 [200] train's binary_logloss: 0.646462 valid's binary_logloss: 0.653161 [300] train's binary_logloss: 0.641483 valid's binary_logloss: 0.650694 [400] train's binary_logloss: 0.637704 valid's binary_logloss: 0.649345 [500] train's binary_logloss: 0.634326 valid's binary_logloss: 0.649458 Early stopping, best iteration is: [458] train's binary_logloss: 0.635679 valid's binary_logloss: 0.648685 regularization_factors, val_score: 0.648669: 25%|##5 | 5/20 [00:05<00:16, 1.09s/it][I 2020-09-27 04:47:39,333] Trial 47 finished with value: 0.6486854676681457 and parameters: {'lambda_l1': 7.797235764236721e-05, 'lambda_l2': 4.7617608412875624e-05}. Best is trial 47 with value: 0.6486854676681457. regularization_factors, val_score: 0.648669: 25%|##5 | 5/20 [00:05<00:16, 1.09s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004751 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656759 valid's binary_logloss: 0.661816 [200] train's binary_logloss: 0.646462 valid's binary_logloss: 0.653161 [300] train's binary_logloss: 0.641483 valid's binary_logloss: 0.650694 [400] train's binary_logloss: 0.6377 valid's binary_logloss: 0.649329 [500] train's binary_logloss: 0.634322 valid's binary_logloss: 0.649442 Early stopping, best iteration is: [458] train's binary_logloss: 0.635675 valid's binary_logloss: 0.648669 regularization_factors, val_score: 0.648669: 30%|### | 6/20 [00:06<00:14, 1.03s/it][I 2020-09-27 04:47:40,228] Trial 48 finished with value: 0.6486693564289175 and parameters: {'lambda_l1': 4.6907860757180705e-07, 'lambda_l2': 0.0001968190366629125}. Best is trial 48 with value: 0.6486693564289175. regularization_factors, val_score: 0.648669: 30%|### | 6/20 [00:06<00:14, 1.03s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000809 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656759 valid's binary_logloss: 0.661816 [200] train's binary_logloss: 0.646461 valid's binary_logloss: 0.653161 [300] train's binary_logloss: 0.641483 valid's binary_logloss: 0.650694 [400] train's binary_logloss: 0.637703 valid's binary_logloss: 0.649345 [500] train's binary_logloss: 0.634325 valid's binary_logloss: 0.649458 Early stopping, best iteration is: [458] train's binary_logloss: 0.635679 valid's binary_logloss: 0.648685 regularization_factors, val_score: 0.648669: 35%|###5 | 7/20 [00:07<00:12, 1.01it/s][I 2020-09-27 04:47:41,143] Trial 49 finished with value: 0.6486853650384133 and parameters: {'lambda_l1': 7.106958303444078e-07, 'lambda_l2': 0.00031503172943699124}. Best is trial 48 with value: 0.6486693564289175. regularization_factors, val_score: 0.648669: 35%|###5 | 7/20 [00:07<00:12, 1.01it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004879 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656759 valid's binary_logloss: 0.661816 [200] train's binary_logloss: 0.646462 valid's binary_logloss: 0.653161 [300] train's binary_logloss: 0.641483 valid's binary_logloss: 0.650694 [400] train's binary_logloss: 0.637704 valid's binary_logloss: 0.649345 [500] train's binary_logloss: 0.634326 valid's binary_logloss: 0.649458 Early stopping, best iteration is: [458] train's binary_logloss: 0.635679 valid's binary_logloss: 0.648685 regularization_factors, val_score: 0.648669: 40%|#### | 8/20 [00:08<00:13, 1.09s/it][I 2020-09-27 04:47:42,447] Trial 50 finished with value: 0.6486854411094597 and parameters: {'lambda_l1': 4.4792617037779606e-07, 'lambda_l2': 0.00040119958572656174}. Best is trial 48 with value: 0.6486693564289175. regularization_factors, val_score: 0.648669: 40%|#### | 8/20 [00:08<00:13, 1.09s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000826 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656759 valid's binary_logloss: 0.661816 [200] train's binary_logloss: 0.646462 valid's binary_logloss: 0.653161 [300] train's binary_logloss: 0.641483 valid's binary_logloss: 0.650694 [400] train's binary_logloss: 0.6377 valid's binary_logloss: 0.649329 [500] train's binary_logloss: 0.634322 valid's binary_logloss: 0.649442 Early stopping, best iteration is: [458] train's binary_logloss: 0.635675 valid's binary_logloss: 0.648669 regularization_factors, val_score: 0.648669: 45%|####5 | 9/20 [00:09<00:11, 1.05s/it][I 2020-09-27 04:47:43,405] Trial 51 finished with value: 0.6486693484397528 and parameters: {'lambda_l1': 3.4434140107307795e-07, 'lambda_l2': 0.00030839964549718207}. Best is trial 51 with value: 0.6486693484397528. regularization_factors, val_score: 0.648669: 45%|####5 | 9/20 [00:09<00:11, 1.05s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000859 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656759 valid's binary_logloss: 0.661816 [200] train's binary_logloss: 0.646462 valid's binary_logloss: 0.653161 [300] train's binary_logloss: 0.641483 valid's binary_logloss: 0.650694 [400] train's binary_logloss: 0.637704 valid's binary_logloss: 0.649345 [500] train's binary_logloss: 0.634326 valid's binary_logloss: 0.649458 Early stopping, best iteration is: [458] train's binary_logloss: 0.635679 valid's binary_logloss: 0.648685 regularization_factors, val_score: 0.648669: 50%|##### | 10/20 [00:10<00:10, 1.01s/it][I 2020-09-27 04:47:44,335] Trial 52 finished with value: 0.6486854548458479 and parameters: {'lambda_l1': 2.9104522149216847e-07, 'lambda_l2': 0.00021910741939920448}. Best is trial 51 with value: 0.6486693484397528. regularization_factors, val_score: 0.648669: 50%|##### | 10/20 [00:10<00:10, 1.01s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009619 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656759 valid's binary_logloss: 0.661816 [200] train's binary_logloss: 0.646462 valid's binary_logloss: 0.653161 [300] train's binary_logloss: 0.641483 valid's binary_logloss: 0.650694 [400] train's binary_logloss: 0.637704 valid's binary_logloss: 0.649345 [500] train's binary_logloss: 0.634326 valid's binary_logloss: 0.649458 Early stopping, best iteration is: [458] train's binary_logloss: 0.635679 valid's binary_logloss: 0.648685 regularization_factors, val_score: 0.648669: 55%|#####5 | 11/20 [00:11<00:09, 1.03s/it][I 2020-09-27 04:47:45,395] Trial 53 finished with value: 0.6486854423114767 and parameters: {'lambda_l1': 1.8051040240033813e-07, 'lambda_l2': 0.0003854090838672508}. Best is trial 51 with value: 0.6486693484397528. regularization_factors, val_score: 0.648669: 55%|#####5 | 11/20 [00:11<00:09, 1.03s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002587 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656759 valid's binary_logloss: 0.661816 [200] train's binary_logloss: 0.646462 valid's binary_logloss: 0.653161 [300] train's binary_logloss: 0.641483 valid's binary_logloss: 0.650694 [400] train's binary_logloss: 0.637704 valid's binary_logloss: 0.649345 [500] train's binary_logloss: 0.634326 valid's binary_logloss: 0.649458 Early stopping, best iteration is: [458] train's binary_logloss: 0.635679 valid's binary_logloss: 0.648685 regularization_factors, val_score: 0.648669: 60%|###### | 12/20 [00:12<00:08, 1.11s/it][I 2020-09-27 04:47:46,692] Trial 54 finished with value: 0.6486853768130912 and parameters: {'lambda_l1': 1.238928167433891e-06, 'lambda_l2': 0.001256677247104136}. Best is trial 51 with value: 0.6486693484397528. regularization_factors, val_score: 0.648669: 60%|###### | 12/20 [00:12<00:08, 1.11s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009674 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656759 valid's binary_logloss: 0.661816 [200] train's binary_logloss: 0.646462 valid's binary_logloss: 0.653161 [300] train's binary_logloss: 0.641484 valid's binary_logloss: 0.650694 [400] train's binary_logloss: 0.637701 valid's binary_logloss: 0.649329 [500] train's binary_logloss: 0.634323 valid's binary_logloss: 0.649442 Early stopping, best iteration is: [458] train's binary_logloss: 0.635676 valid's binary_logloss: 0.648669 regularization_factors, val_score: 0.648669: 65%|######5 | 13/20 [00:13<00:07, 1.05s/it][I 2020-09-27 04:47:47,595] Trial 55 finished with value: 0.6486691459174234 and parameters: {'lambda_l1': 4.463187899572695e-06, 'lambda_l2': 0.0031440113191708815}. Best is trial 55 with value: 0.6486691459174234. regularization_factors, val_score: 0.648669: 65%|######5 | 13/20 [00:13<00:07, 1.05s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004983 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656759 valid's binary_logloss: 0.661816 [200] train's binary_logloss: 0.646463 valid's binary_logloss: 0.653161 [300] train's binary_logloss: 0.641496 valid's binary_logloss: 0.650455 [400] train's binary_logloss: 0.637726 valid's binary_logloss: 0.649177 [500] train's binary_logloss: 0.634323 valid's binary_logloss: 0.649235 Early stopping, best iteration is: [458] train's binary_logloss: 0.635717 valid's binary_logloss: 0.648475 regularization_factors, val_score: 0.648475: 70%|####### | 14/20 [00:14<00:06, 1.00s/it][I 2020-09-27 04:47:48,497] Trial 56 finished with value: 0.6484745400508054 and parameters: {'lambda_l1': 9.72103845454671e-06, 'lambda_l2': 0.011377674391625028}. Best is trial 56 with value: 0.6484745400508054. regularization_factors, val_score: 0.648475: 70%|####### | 14/20 [00:14<00:06, 1.00s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001108 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656804 valid's binary_logloss: 0.661432 [200] train's binary_logloss: 0.646512 valid's binary_logloss: 0.653055 [300] train's binary_logloss: 0.641411 valid's binary_logloss: 0.650947 [400] train's binary_logloss: 0.637723 valid's binary_logloss: 0.649987 [500] train's binary_logloss: 0.634326 valid's binary_logloss: 0.650052 Early stopping, best iteration is: [464] train's binary_logloss: 0.635548 valid's binary_logloss: 0.649488 regularization_factors, val_score: 0.648475: 75%|#######5 | 15/20 [00:15<00:05, 1.08s/it][I 2020-09-27 04:47:49,768] Trial 57 finished with value: 0.6494877361955586 and parameters: {'lambda_l1': 1.3431278103969494e-05, 'lambda_l2': 0.02447854418046466}. Best is trial 56 with value: 0.6484745400508054. regularization_factors, val_score: 0.648475: 75%|#######5 | 15/20 [00:15<00:05, 1.08s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.015062 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656759 valid's binary_logloss: 0.661816 [200] train's binary_logloss: 0.646462 valid's binary_logloss: 0.653161 [300] train's binary_logloss: 0.641495 valid's binary_logloss: 0.650455 [400] train's binary_logloss: 0.637725 valid's binary_logloss: 0.649177 [500] train's binary_logloss: 0.634322 valid's binary_logloss: 0.649235 Early stopping, best iteration is: [458] train's binary_logloss: 0.635715 valid's binary_logloss: 0.648475 regularization_factors, val_score: 0.648475: 80%|######## | 16/20 [00:16<00:04, 1.08s/it][I 2020-09-27 04:47:50,844] Trial 58 finished with value: 0.6484747613620405 and parameters: {'lambda_l1': 5.558047158433351e-06, 'lambda_l2': 0.007068676788026479}. Best is trial 56 with value: 0.6484745400508054. regularization_factors, val_score: 0.648475: 80%|######## | 16/20 [00:16<00:04, 1.08s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000877 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656759 valid's binary_logloss: 0.661817 [200] train's binary_logloss: 0.646463 valid's binary_logloss: 0.653162 [300] train's binary_logloss: 0.641497 valid's binary_logloss: 0.650455 [400] train's binary_logloss: 0.637728 valid's binary_logloss: 0.649176 [500] train's binary_logloss: 0.634326 valid's binary_logloss: 0.649234 Early stopping, best iteration is: [458] train's binary_logloss: 0.635719 valid's binary_logloss: 0.648474 regularization_factors, val_score: 0.648474: 85%|########5 | 17/20 [00:17<00:03, 1.09s/it][I 2020-09-27 04:47:51,940] Trial 59 finished with value: 0.6484741879925516 and parameters: {'lambda_l1': 1.1402866509712536e-05, 'lambda_l2': 0.018245531549826995}. Best is trial 59 with value: 0.6484741879925516. regularization_factors, val_score: 0.648474: 85%|########5 | 17/20 [00:17<00:03, 1.09s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004790 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656759 valid's binary_logloss: 0.661817 [200] train's binary_logloss: 0.646463 valid's binary_logloss: 0.653162 [300] train's binary_logloss: 0.641497 valid's binary_logloss: 0.650455 [400] train's binary_logloss: 0.637728 valid's binary_logloss: 0.649176 [500] train's binary_logloss: 0.634326 valid's binary_logloss: 0.649234 Early stopping, best iteration is: [458] train's binary_logloss: 0.635719 valid's binary_logloss: 0.648474 regularization_factors, val_score: 0.648474: 90%|######### | 18/20 [00:18<00:02, 1.03s/it][I 2020-09-27 04:47:52,825] Trial 60 finished with value: 0.6484741937796707 and parameters: {'lambda_l1': 1.9414476880011605e-05, 'lambda_l2': 0.018137780474702055}. Best is trial 59 with value: 0.6484741879925516. regularization_factors, val_score: 0.648474: 90%|######### | 18/20 [00:18<00:02, 1.03s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004872 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656759 valid's binary_logloss: 0.661816 [200] train's binary_logloss: 0.646463 valid's binary_logloss: 0.653161 [300] train's binary_logloss: 0.641496 valid's binary_logloss: 0.650455 [400] train's binary_logloss: 0.637727 valid's binary_logloss: 0.649177 [500] train's binary_logloss: 0.634324 valid's binary_logloss: 0.649235 Early stopping, best iteration is: [458] train's binary_logloss: 0.635717 valid's binary_logloss: 0.648474 regularization_factors, val_score: 0.648474: 95%|#########5| 19/20 [00:20<00:01, 1.14s/it][I 2020-09-27 04:47:54,216] Trial 61 finished with value: 0.6484744650230337 and parameters: {'lambda_l1': 1.6059055968360456e-05, 'lambda_l2': 0.012844094468385317}. Best is trial 59 with value: 0.6484741879925516. regularization_factors, val_score: 0.648474: 95%|#########5| 19/20 [00:20<00:01, 1.14s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004738 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.65676 valid's binary_logloss: 0.661817 [200] train's binary_logloss: 0.646463 valid's binary_logloss: 0.653162 [300] train's binary_logloss: 0.641498 valid's binary_logloss: 0.650454 [400] train's binary_logloss: 0.637729 valid's binary_logloss: 0.649176 [500] train's binary_logloss: 0.634329 valid's binary_logloss: 0.649256 Early stopping, best iteration is: [458] train's binary_logloss: 0.63572 valid's binary_logloss: 0.648474 regularization_factors, val_score: 0.648474: 100%|##########| 20/20 [00:21<00:00, 1.09s/it][I 2020-09-27 04:47:55,185] Trial 62 finished with value: 0.6484739197241262 and parameters: {'lambda_l1': 1.4740300654925883e-05, 'lambda_l2': 0.02349270671135121}. Best is trial 62 with value: 0.6484739197241262. regularization_factors, val_score: 0.648474: 100%|##########| 20/20 [00:21<00:00, 1.06s/it] min_data_in_leaf, val_score: 0.648474: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000891 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.65676 valid's binary_logloss: 0.661817 [200] train's binary_logloss: 0.646525 valid's binary_logloss: 0.653595 [300] train's binary_logloss: 0.641846 valid's binary_logloss: 0.650251 [400] train's binary_logloss: 0.638553 valid's binary_logloss: 0.649577 [500] train's binary_logloss: 0.63552 valid's binary_logloss: 0.64935 Early stopping, best iteration is: [464] train's binary_logloss: 0.636644 valid's binary_logloss: 0.648879 min_data_in_leaf, val_score: 0.648474: 20%|## | 1/5 [00:01<00:04, 1.01s/it][I 2020-09-27 04:47:56,208] Trial 63 finished with value: 0.648879477038123 and parameters: {'min_child_samples': 100}. Best is trial 63 with value: 0.648879477038123. min_data_in_leaf, val_score: 0.648474: 20%|## | 1/5 [00:01<00:04, 1.01s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010338 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.65676 valid's binary_logloss: 0.661817 [200] train's binary_logloss: 0.646463 valid's binary_logloss: 0.653162 [300] train's binary_logloss: 0.6415 valid's binary_logloss: 0.650452 [400] train's binary_logloss: 0.637791 valid's binary_logloss: 0.649375 [500] train's binary_logloss: 0.634508 valid's binary_logloss: 0.649431 Early stopping, best iteration is: [451] train's binary_logloss: 0.636075 valid's binary_logloss: 0.648808 min_data_in_leaf, val_score: 0.648474: 40%|#### | 2/5 [00:02<00:03, 1.03s/it][I 2020-09-27 04:47:57,286] Trial 64 finished with value: 0.6488082911033146 and parameters: {'min_child_samples': 25}. Best is trial 64 with value: 0.6488082911033146. min_data_in_leaf, val_score: 0.648474: 40%|#### | 2/5 [00:02<00:03, 1.03s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.015161 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.65676 valid's binary_logloss: 0.661817 [200] train's binary_logloss: 0.646493 valid's binary_logloss: 0.65325 [300] train's binary_logloss: 0.641497 valid's binary_logloss: 0.650822 [400] train's binary_logloss: 0.637788 valid's binary_logloss: 0.649913 [500] train's binary_logloss: 0.634246 valid's binary_logloss: 0.649666 Early stopping, best iteration is: [468] train's binary_logloss: 0.635328 valid's binary_logloss: 0.649123 min_data_in_leaf, val_score: 0.648474: 60%|###### | 3/5 [00:03<00:02, 1.10s/it][I 2020-09-27 04:47:58,535] Trial 65 finished with value: 0.649122937898972 and parameters: {'min_child_samples': 10}. Best is trial 64 with value: 0.6488082911033146. min_data_in_leaf, val_score: 0.648474: 60%|###### | 3/5 [00:03<00:02, 1.10s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001909 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.65676 valid's binary_logloss: 0.661817 [200] train's binary_logloss: 0.64648 valid's binary_logloss: 0.653437 [300] train's binary_logloss: 0.641645 valid's binary_logloss: 0.650437 [400] train's binary_logloss: 0.638049 valid's binary_logloss: 0.649312 [500] train's binary_logloss: 0.634894 valid's binary_logloss: 0.649312 Early stopping, best iteration is: [447] train's binary_logloss: 0.63654 valid's binary_logloss: 0.648959 min_data_in_leaf, val_score: 0.648474: 80%|######## | 4/5 [00:04<00:01, 1.04s/it][I 2020-09-27 04:47:59,458] Trial 66 finished with value: 0.6489591737909831 and parameters: {'min_child_samples': 50}. Best is trial 64 with value: 0.6488082911033146. min_data_in_leaf, val_score: 0.648474: 80%|######## | 4/5 [00:04<00:01, 1.04s/it][LightGBM] [Info] Number of positive: 12855, number of negative: 13145 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001493 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4242 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494423 -> initscore=-0.022309 [LightGBM] [Info] Start training from score -0.022309 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.65676 valid's binary_logloss: 0.661817 [200] train's binary_logloss: 0.646493 valid's binary_logloss: 0.65325 [300] train's binary_logloss: 0.64151 valid's binary_logloss: 0.65102 [400] train's binary_logloss: 0.637718 valid's binary_logloss: 0.650205 [500] train's binary_logloss: 0.634224 valid's binary_logloss: 0.649812 Early stopping, best iteration is: [459] train's binary_logloss: 0.635631 valid's binary_logloss: 0.649459 min_data_in_leaf, val_score: 0.648474: 100%|##########| 5/5 [00:05<00:00, 1.02s/it][I 2020-09-27 04:48:00,406] Trial 67 finished with value: 0.6494585505180116 and parameters: {'min_child_samples': 5}. Best is trial 64 with value: 0.6488082911033146. min_data_in_leaf, val_score: 0.648474: 100%|##########| 5/5 [00:05<00:00, 1.04s/it] Fold : 9 [I 2020-09-27 04:48:00,469] A new study created in memory with name: no-name-7b517b46-d774-4255-8991-4b95254b4cf4 feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000996 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.571415 valid's binary_logloss: 0.664205 Early stopping, best iteration is: [56] train's binary_logloss: 0.602615 valid's binary_logloss: 0.661775 feature_fraction, val_score: 0.661775: 14%|#4 | 1/7 [00:00<00:05, 1.12it/s][I 2020-09-27 04:48:01,375] Trial 0 finished with value: 0.6617754129584598 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 0 with value: 0.6617754129584598. feature_fraction, val_score: 0.661775: 14%|#4 | 1/7 [00:00<00:05, 1.12it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001704 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.577907 valid's binary_logloss: 0.66332 Early stopping, best iteration is: [71] train's binary_logloss: 0.596884 valid's binary_logloss: 0.662116 feature_fraction, val_score: 0.661775: 29%|##8 | 2/7 [00:02<00:04, 1.04it/s][I 2020-09-27 04:48:02,518] Trial 1 finished with value: 0.6621160574043524 and parameters: {'feature_fraction': 0.5}. Best is trial 0 with value: 0.6617754129584598. feature_fraction, val_score: 0.661775: 29%|##8 | 2/7 [00:02<00:04, 1.04it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000446 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.581403 valid's binary_logloss: 0.664323 Early stopping, best iteration is: [69] train's binary_logloss: 0.601689 valid's binary_logloss: 0.663133 feature_fraction, val_score: 0.661775: 43%|####2 | 3/7 [00:02<00:03, 1.21it/s][I 2020-09-27 04:48:03,026] Trial 2 finished with value: 0.6631333345540273 and parameters: {'feature_fraction': 0.4}. Best is trial 0 with value: 0.6617754129584598. feature_fraction, val_score: 0.661775: 43%|####2 | 3/7 [00:02<00:03, 1.21it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000926 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.569848 valid's binary_logloss: 0.664707 Early stopping, best iteration is: [67] train's binary_logloss: 0.59321 valid's binary_logloss: 0.662756 feature_fraction, val_score: 0.661775: 57%|#####7 | 4/7 [00:03<00:02, 1.35it/s][I 2020-09-27 04:48:03,555] Trial 3 finished with value: 0.6627562719479632 and parameters: {'feature_fraction': 1.0}. Best is trial 0 with value: 0.6617754129584598. feature_fraction, val_score: 0.661775: 57%|#####7 | 4/7 [00:03<00:02, 1.35it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000814 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.573306 valid's binary_logloss: 0.661877 [200] train's binary_logloss: 0.517495 valid's binary_logloss: 0.664844 Early stopping, best iteration is: [113] train's binary_logloss: 0.565412 valid's binary_logloss: 0.660855 feature_fraction, val_score: 0.660855: 71%|#######1 | 5/7 [00:03<00:01, 1.42it/s][I 2020-09-27 04:48:04,180] Trial 4 finished with value: 0.6608549078376256 and parameters: {'feature_fraction': 0.8}. Best is trial 4 with value: 0.6608549078376256. feature_fraction, val_score: 0.660855: 71%|#######1 | 5/7 [00:03<00:01, 1.42it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004908 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.574579 valid's binary_logloss: 0.662991 Early stopping, best iteration is: [54] train's binary_logloss: 0.606825 valid's binary_logloss: 0.661109 feature_fraction, val_score: 0.660855: 86%|########5 | 6/7 [00:04<00:00, 1.58it/s][I 2020-09-27 04:48:04,649] Trial 5 finished with value: 0.66110942300072 and parameters: {'feature_fraction': 0.7}. Best is trial 4 with value: 0.6608549078376256. feature_fraction, val_score: 0.660855: 86%|########5 | 6/7 [00:04<00:00, 1.58it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004781 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.575121 valid's binary_logloss: 0.665199 Early stopping, best iteration is: [58] train's binary_logloss: 0.604585 valid's binary_logloss: 0.663434 feature_fraction, val_score: 0.660855: 100%|##########| 7/7 [00:04<00:00, 1.58it/s][I 2020-09-27 04:48:05,278] Trial 6 finished with value: 0.6634341152776219 and parameters: {'feature_fraction': 0.6}. Best is trial 4 with value: 0.6608549078376256. feature_fraction, val_score: 0.660855: 100%|##########| 7/7 [00:04<00:00, 1.46it/s] num_leaves, val_score: 0.660855: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004289 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.573306 valid's binary_logloss: 0.661877 [200] train's binary_logloss: 0.517495 valid's binary_logloss: 0.664844 Early stopping, best iteration is: [113] train's binary_logloss: 0.565412 valid's binary_logloss: 0.660855 num_leaves, val_score: 0.660855: 5%|5 | 1/20 [00:01<00:37, 1.99s/it][I 2020-09-27 04:48:07,289] Trial 7 finished with value: 0.6608549078376256 and parameters: {'num_leaves': 31}. Best is trial 7 with value: 0.6608549078376256. num_leaves, val_score: 0.660855: 5%|5 | 1/20 [00:02<00:37, 1.99s/it][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.009560 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.291282 valid's binary_logloss: 0.683399 Early stopping, best iteration is: [25] train's binary_logloss: 0.515751 valid's binary_logloss: 0.662419 num_leaves, val_score: 0.660855: 10%|# | 2/20 [00:03<00:32, 1.79s/it][I 2020-09-27 04:48:08,615] Trial 8 finished with value: 0.6624189569421138 and parameters: {'num_leaves': 227}. Best is trial 7 with value: 0.6608549078376256. num_leaves, val_score: 0.660855: 10%|# | 2/20 [00:03<00:32, 1.79s/it][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000898 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.633793 valid's binary_logloss: 0.66169 [200] train's binary_logloss: 0.616176 valid's binary_logloss: 0.662116 Early stopping, best iteration is: [127] train's binary_logloss: 0.628472 valid's binary_logloss: 0.661027 num_leaves, val_score: 0.660855: 15%|#5 | 3/20 [00:03<00:24, 1.41s/it][I 2020-09-27 04:48:09,146] Trial 9 finished with value: 0.6610265822259738 and parameters: {'num_leaves': 8}. Best is trial 7 with value: 0.6608549078376256. num_leaves, val_score: 0.660855: 15%|#5 | 3/20 [00:03<00:24, 1.41s/it][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000937 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.277955 valid's binary_logloss: 0.68283 Early stopping, best iteration is: [29] train's binary_logloss: 0.489512 valid's binary_logloss: 0.667966 num_leaves, val_score: 0.660855: 20%|## | 4/20 [00:05<00:24, 1.52s/it][I 2020-09-27 04:48:10,921] Trial 10 finished with value: 0.6679657755747807 and parameters: {'num_leaves': 241}. Best is trial 7 with value: 0.6608549078376256. num_leaves, val_score: 0.660855: 20%|## | 4/20 [00:05<00:24, 1.52s/it][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000991 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.586674 valid's binary_logloss: 0.66365 Early stopping, best iteration is: [74] train's binary_logloss: 0.601053 valid's binary_logloss: 0.662634 num_leaves, val_score: 0.660855: 25%|##5 | 5/20 [00:06<00:18, 1.21s/it][I 2020-09-27 04:48:11,417] Trial 11 finished with value: 0.6626335832631651 and parameters: {'num_leaves': 25}. Best is trial 7 with value: 0.6608549078376256. num_leaves, val_score: 0.660855: 25%|##5 | 5/20 [00:06<00:18, 1.21s/it][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000938 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.437791 valid's binary_logloss: 0.672762 Early stopping, best iteration is: [35] train's binary_logloss: 0.555775 valid's binary_logloss: 0.664193 num_leaves, val_score: 0.660855: 30%|### | 6/20 [00:06<00:15, 1.07s/it][I 2020-09-27 04:48:12,160] Trial 12 finished with value: 0.664192632015902 and parameters: {'num_leaves': 104}. Best is trial 7 with value: 0.6608549078376256. num_leaves, val_score: 0.660855: 30%|### | 6/20 [00:06<00:15, 1.07s/it][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000963 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.458686 valid's binary_logloss: 0.668395 Early stopping, best iteration is: [40] train's binary_logloss: 0.556167 valid's binary_logloss: 0.664721 num_leaves, val_score: 0.660855: 35%|###5 | 7/20 [00:07<00:12, 1.03it/s][I 2020-09-27 04:48:12,890] Trial 13 finished with value: 0.6647214922013416 and parameters: {'num_leaves': 90}. Best is trial 7 with value: 0.6608549078376256. num_leaves, val_score: 0.660855: 35%|###5 | 7/20 [00:07<00:12, 1.03it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009252 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.350518 valid's binary_logloss: 0.684373 Early stopping, best iteration is: [22] train's binary_logloss: 0.558627 valid's binary_logloss: 0.66659 num_leaves, val_score: 0.660855: 40%|#### | 8/20 [00:08<00:11, 1.07it/s][I 2020-09-27 04:48:13,742] Trial 14 finished with value: 0.6665899695534304 and parameters: {'num_leaves': 167}. Best is trial 7 with value: 0.6608549078376256. num_leaves, val_score: 0.660855: 40%|#### | 8/20 [00:08<00:11, 1.07it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004731 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.532376 valid's binary_logloss: 0.669923 Early stopping, best iteration is: [40] train's binary_logloss: 0.596466 valid's binary_logloss: 0.665493 num_leaves, val_score: 0.660855: 45%|####5 | 9/20 [00:09<00:10, 1.04it/s][I 2020-09-27 04:48:14,759] Trial 15 finished with value: 0.665493198325232 and parameters: {'num_leaves': 50}. Best is trial 7 with value: 0.6608549078376256. num_leaves, val_score: 0.660855: 45%|####5 | 9/20 [00:09<00:10, 1.04it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000797 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.350518 valid's binary_logloss: 0.684373 Early stopping, best iteration is: [22] train's binary_logloss: 0.558627 valid's binary_logloss: 0.66659 num_leaves, val_score: 0.660855: 50%|##### | 10/20 [00:10<00:09, 1.04it/s][I 2020-09-27 04:48:15,722] Trial 16 finished with value: 0.6665899695534304 and parameters: {'num_leaves': 167}. Best is trial 7 with value: 0.6608549078376256. num_leaves, val_score: 0.660855: 50%|##### | 10/20 [00:10<00:09, 1.04it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004981 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.512177 valid's binary_logloss: 0.666645 Early stopping, best iteration is: [29] train's binary_logloss: 0.60465 valid's binary_logloss: 0.662506 num_leaves, val_score: 0.660855: 55%|#####5 | 11/20 [00:10<00:07, 1.20it/s][I 2020-09-27 04:48:16,266] Trial 17 finished with value: 0.6625059680700698 and parameters: {'num_leaves': 60}. Best is trial 7 with value: 0.6608549078376256. num_leaves, val_score: 0.660855: 55%|#####5 | 11/20 [00:10<00:07, 1.20it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001403 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650402 valid's binary_logloss: 0.665088 [200] train's binary_logloss: 0.639194 valid's binary_logloss: 0.660755 [300] train's binary_logloss: 0.632319 valid's binary_logloss: 0.66028 [400] train's binary_logloss: 0.626788 valid's binary_logloss: 0.660425 Early stopping, best iteration is: [360] train's binary_logloss: 0.628825 valid's binary_logloss: 0.659947 num_leaves, val_score: 0.659947: 60%|###### | 12/20 [00:11<00:06, 1.22it/s][I 2020-09-27 04:48:17,046] Trial 18 finished with value: 0.659946568502427 and parameters: {'num_leaves': 4}. Best is trial 18 with value: 0.659946568502427. num_leaves, val_score: 0.659947: 60%|###### | 12/20 [00:11<00:06, 1.22it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001058 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.637047 valid's binary_logloss: 0.663485 [200] train's binary_logloss: 0.621339 valid's binary_logloss: 0.662464 [300] train's binary_logloss: 0.608631 valid's binary_logloss: 0.662754 Early stopping, best iteration is: [218] train's binary_logloss: 0.618989 valid's binary_logloss: 0.661824 num_leaves, val_score: 0.659947: 65%|######5 | 13/20 [00:12<00:05, 1.31it/s][I 2020-09-27 04:48:17,672] Trial 19 finished with value: 0.6618235788394593 and parameters: {'num_leaves': 7}. Best is trial 18 with value: 0.659946568502427. num_leaves, val_score: 0.659947: 65%|######5 | 13/20 [00:12<00:05, 1.31it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.006735 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657096 valid's binary_logloss: 0.667367 [200] train's binary_logloss: 0.646565 valid's binary_logloss: 0.661837 [300] train's binary_logloss: 0.641199 valid's binary_logloss: 0.660556 [400] train's binary_logloss: 0.637372 valid's binary_logloss: 0.660306 Early stopping, best iteration is: [357] train's binary_logloss: 0.638883 valid's binary_logloss: 0.659984 num_leaves, val_score: 0.659947: 70%|####### | 14/20 [00:13<00:05, 1.13it/s][I 2020-09-27 04:48:18,847] Trial 20 finished with value: 0.6599838058476127 and parameters: {'num_leaves': 3}. Best is trial 18 with value: 0.659946568502427. num_leaves, val_score: 0.659947: 70%|####### | 14/20 [00:13<00:05, 1.13it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001022 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.645096 valid's binary_logloss: 0.662705 [200] train's binary_logloss: 0.632688 valid's binary_logloss: 0.660038 Early stopping, best iteration is: [193] train's binary_logloss: 0.633406 valid's binary_logloss: 0.660001 num_leaves, val_score: 0.659947: 75%|#######5 | 15/20 [00:14<00:03, 1.26it/s][I 2020-09-27 04:48:19,418] Trial 21 finished with value: 0.6600011118675012 and parameters: {'num_leaves': 5}. Best is trial 18 with value: 0.659946568502427. num_leaves, val_score: 0.659947: 75%|#######5 | 15/20 [00:14<00:03, 1.26it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001022 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.629941 valid's binary_logloss: 0.66087 [200] train's binary_logloss: 0.610687 valid's binary_logloss: 0.662744 Early stopping, best iteration is: [125] train's binary_logloss: 0.624522 valid's binary_logloss: 0.66051 num_leaves, val_score: 0.659947: 80%|######## | 16/20 [00:14<00:02, 1.38it/s][I 2020-09-27 04:48:19,995] Trial 22 finished with value: 0.6605096965777126 and parameters: {'num_leaves': 9}. Best is trial 18 with value: 0.659946568502427. num_leaves, val_score: 0.659947: 80%|######## | 16/20 [00:14<00:02, 1.38it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.006817 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.512177 valid's binary_logloss: 0.666645 Early stopping, best iteration is: [29] train's binary_logloss: 0.60465 valid's binary_logloss: 0.662506 num_leaves, val_score: 0.659947: 85%|########5 | 17/20 [00:15<00:02, 1.48it/s][I 2020-09-27 04:48:20,555] Trial 23 finished with value: 0.6625059680700698 and parameters: {'num_leaves': 60}. Best is trial 18 with value: 0.659946568502427. num_leaves, val_score: 0.659947: 85%|########5 | 17/20 [00:15<00:02, 1.48it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001075 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650402 valid's binary_logloss: 0.665088 [200] train's binary_logloss: 0.639194 valid's binary_logloss: 0.660755 [300] train's binary_logloss: 0.632319 valid's binary_logloss: 0.66028 [400] train's binary_logloss: 0.626788 valid's binary_logloss: 0.660425 Early stopping, best iteration is: [360] train's binary_logloss: 0.628825 valid's binary_logloss: 0.659947 num_leaves, val_score: 0.659947: 90%|######### | 18/20 [00:16<00:01, 1.40it/s][I 2020-09-27 04:48:21,359] Trial 24 finished with value: 0.6599465685024268 and parameters: {'num_leaves': 4}. Best is trial 24 with value: 0.6599465685024268. num_leaves, val_score: 0.659947: 90%|######### | 18/20 [00:16<00:01, 1.40it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000833 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.569877 valid's binary_logloss: 0.667063 Early stopping, best iteration is: [43] train's binary_logloss: 0.613515 valid's binary_logloss: 0.663355 num_leaves, val_score: 0.659947: 95%|#########5| 19/20 [00:17<00:00, 1.27it/s][I 2020-09-27 04:48:22,313] Trial 25 finished with value: 0.6633553868179578 and parameters: {'num_leaves': 32}. Best is trial 24 with value: 0.6599465685024268. num_leaves, val_score: 0.659947: 95%|#########5| 19/20 [00:17<00:00, 1.27it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000762 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.466555 valid's binary_logloss: 0.674499 Early stopping, best iteration is: [32] train's binary_logloss: 0.577932 valid's binary_logloss: 0.665241 num_leaves, val_score: 0.659947: 100%|##########| 20/20 [00:17<00:00, 1.30it/s][I 2020-09-27 04:48:23,045] Trial 26 finished with value: 0.6652408810612898 and parameters: {'num_leaves': 85}. Best is trial 24 with value: 0.6599465685024268. num_leaves, val_score: 0.659947: 100%|##########| 20/20 [00:17<00:00, 1.13it/s] bagging, val_score: 0.659947: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000831 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.649887 valid's binary_logloss: 0.663942 [200] train's binary_logloss: 0.639122 valid's binary_logloss: 0.660082 [300] train's binary_logloss: 0.632108 valid's binary_logloss: 0.660084 Early stopping, best iteration is: [239] train's binary_logloss: 0.636191 valid's binary_logloss: 0.659231 bagging, val_score: 0.659231: 10%|# | 1/10 [00:00<00:06, 1.48it/s][I 2020-09-27 04:48:23,737] Trial 27 finished with value: 0.6592308295470516 and parameters: {'bagging_fraction': 0.8273856061488931, 'bagging_freq': 3}. Best is trial 27 with value: 0.6592308295470516. bagging, val_score: 0.659231: 10%|# | 1/10 [00:00<00:06, 1.48it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004809 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.649764 valid's binary_logloss: 0.663649 [200] train's binary_logloss: 0.638988 valid's binary_logloss: 0.65972 [300] train's binary_logloss: 0.632186 valid's binary_logloss: 0.659366 [400] train's binary_logloss: 0.626354 valid's binary_logloss: 0.659639 Early stopping, best iteration is: [322] train's binary_logloss: 0.630893 valid's binary_logloss: 0.659046 bagging, val_score: 0.659046: 20%|## | 2/10 [00:01<00:05, 1.42it/s][I 2020-09-27 04:48:24,511] Trial 28 finished with value: 0.659045827636086 and parameters: {'bagging_fraction': 0.8505997166142588, 'bagging_freq': 3}. Best is trial 28 with value: 0.659045827636086. bagging, val_score: 0.659046: 20%|## | 2/10 [00:01<00:05, 1.42it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.006005 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448 [200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404 [300] train's binary_logloss: 0.631879 valid's binary_logloss: 0.657967 Early stopping, best iteration is: [281] train's binary_logloss: 0.633137 valid's binary_logloss: 0.657825 bagging, val_score: 0.657825: 30%|### | 3/10 [00:02<00:05, 1.40it/s][I 2020-09-27 04:48:25,250] Trial 29 finished with value: 0.6578254229578112 and parameters: {'bagging_fraction': 0.8638496845155332, 'bagging_freq': 3}. Best is trial 29 with value: 0.6578254229578112. bagging, val_score: 0.657825: 30%|### | 3/10 [00:02<00:05, 1.40it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001015 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.649823 valid's binary_logloss: 0.663663 [200] train's binary_logloss: 0.638792 valid's binary_logloss: 0.659288 [300] train's binary_logloss: 0.631889 valid's binary_logloss: 0.658487 Early stopping, best iteration is: [241] train's binary_logloss: 0.635846 valid's binary_logloss: 0.658154 bagging, val_score: 0.657825: 40%|#### | 4/10 [00:03<00:04, 1.28it/s][I 2020-09-27 04:48:26,194] Trial 30 finished with value: 0.6581539755446953 and parameters: {'bagging_fraction': 0.8557672308630124, 'bagging_freq': 3}. Best is trial 29 with value: 0.6578254229578112. bagging, val_score: 0.657825: 40%|#### | 4/10 [00:03<00:04, 1.28it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002755 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650081 valid's binary_logloss: 0.663971 [200] train's binary_logloss: 0.639132 valid's binary_logloss: 0.660066 [300] train's binary_logloss: 0.632409 valid's binary_logloss: 0.659247 [400] train's binary_logloss: 0.626618 valid's binary_logloss: 0.65931 Early stopping, best iteration is: [341] train's binary_logloss: 0.630063 valid's binary_logloss: 0.659011 bagging, val_score: 0.657825: 50%|##### | 5/10 [00:04<00:04, 1.12it/s][I 2020-09-27 04:48:27,332] Trial 31 finished with value: 0.6590114175208794 and parameters: {'bagging_fraction': 0.8641735815144664, 'bagging_freq': 3}. Best is trial 29 with value: 0.6578254229578112. bagging, val_score: 0.657825: 50%|##### | 5/10 [00:04<00:04, 1.12it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000894 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.64988 valid's binary_logloss: 0.664431 [200] train's binary_logloss: 0.639026 valid's binary_logloss: 0.660245 [300] train's binary_logloss: 0.632354 valid's binary_logloss: 0.659402 [400] train's binary_logloss: 0.626336 valid's binary_logloss: 0.659571 Early stopping, best iteration is: [322] train's binary_logloss: 0.630965 valid's binary_logloss: 0.659137 bagging, val_score: 0.657825: 60%|###### | 6/10 [00:05<00:03, 1.17it/s][I 2020-09-27 04:48:28,113] Trial 32 finished with value: 0.6591365814377496 and parameters: {'bagging_fraction': 0.8683729064753997, 'bagging_freq': 3}. Best is trial 29 with value: 0.6578254229578112. bagging, val_score: 0.657825: 60%|###### | 6/10 [00:05<00:03, 1.17it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.006198 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.649852 valid's binary_logloss: 0.663933 [200] train's binary_logloss: 0.639035 valid's binary_logloss: 0.65961 [300] train's binary_logloss: 0.632241 valid's binary_logloss: 0.658575 [400] train's binary_logloss: 0.62642 valid's binary_logloss: 0.658789 Early stopping, best iteration is: [362] train's binary_logloss: 0.628634 valid's binary_logloss: 0.658088 bagging, val_score: 0.657825: 70%|####### | 7/10 [00:05<00:02, 1.16it/s][I 2020-09-27 04:48:28,978] Trial 33 finished with value: 0.6580884313848213 and parameters: {'bagging_fraction': 0.8714478568851503, 'bagging_freq': 3}. Best is trial 29 with value: 0.6578254229578112. bagging, val_score: 0.657825: 70%|####### | 7/10 [00:05<00:02, 1.16it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000833 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.649854 valid's binary_logloss: 0.66428 [200] train's binary_logloss: 0.638784 valid's binary_logloss: 0.660031 [300] train's binary_logloss: 0.631851 valid's binary_logloss: 0.659733 Early stopping, best iteration is: [259] train's binary_logloss: 0.634487 valid's binary_logloss: 0.659495 bagging, val_score: 0.657825: 80%|######## | 8/10 [00:06<00:01, 1.20it/s][I 2020-09-27 04:48:29,747] Trial 34 finished with value: 0.6594947433812736 and parameters: {'bagging_fraction': 0.9161669099599079, 'bagging_freq': 3}. Best is trial 29 with value: 0.6578254229578112. bagging, val_score: 0.657825: 80%|######## | 8/10 [00:06<00:01, 1.20it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.003234 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.649122 valid's binary_logloss: 0.66312 [200] train's binary_logloss: 0.638562 valid's binary_logloss: 0.660896 Early stopping, best iteration is: [170] train's binary_logloss: 0.640905 valid's binary_logloss: 0.660106 bagging, val_score: 0.657825: 90%|######### | 9/10 [00:07<00:00, 1.18it/s][I 2020-09-27 04:48:30,634] Trial 35 finished with value: 0.6601056363582504 and parameters: {'bagging_fraction': 0.5253666376896976, 'bagging_freq': 2}. Best is trial 29 with value: 0.6578254229578112. bagging, val_score: 0.657825: 90%|######### | 9/10 [00:07<00:00, 1.18it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000924 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.64967 valid's binary_logloss: 0.663468 [200] train's binary_logloss: 0.638905 valid's binary_logloss: 0.660271 [300] train's binary_logloss: 0.632375 valid's binary_logloss: 0.660007 [400] train's binary_logloss: 0.626598 valid's binary_logloss: 0.659422 [500] train's binary_logloss: 0.621008 valid's binary_logloss: 0.658741 Early stopping, best iteration is: [478] train's binary_logloss: 0.622298 valid's binary_logloss: 0.658486 bagging, val_score: 0.657825: 100%|##########| 10/10 [00:08<00:00, 1.12it/s][I 2020-09-27 04:48:31,633] Trial 36 finished with value: 0.658485788443909 and parameters: {'bagging_fraction': 0.7531472629337493, 'bagging_freq': 6}. Best is trial 29 with value: 0.6578254229578112. bagging, val_score: 0.657825: 100%|##########| 10/10 [00:08<00:00, 1.17it/s] feature_fraction_stage2, val_score: 0.657825: 0%| | 0/6 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000919 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.649945 valid's binary_logloss: 0.663854 [200] train's binary_logloss: 0.638845 valid's binary_logloss: 0.659895 [300] train's binary_logloss: 0.631902 valid's binary_logloss: 0.659391 [400] train's binary_logloss: 0.626123 valid's binary_logloss: 0.659153 Early stopping, best iteration is: [373] train's binary_logloss: 0.627702 valid's binary_logloss: 0.658836 feature_fraction_stage2, val_score: 0.657825: 17%|#6 | 1/6 [00:00<00:04, 1.09it/s][I 2020-09-27 04:48:32,563] Trial 37 finished with value: 0.6588361247453068 and parameters: {'feature_fraction': 0.8480000000000001}. Best is trial 37 with value: 0.6588361247453068. feature_fraction_stage2, val_score: 0.657825: 17%|#6 | 1/6 [00:00<00:04, 1.09it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000843 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448 [200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404 [300] train's binary_logloss: 0.631879 valid's binary_logloss: 0.657967 Early stopping, best iteration is: [281] train's binary_logloss: 0.633137 valid's binary_logloss: 0.657825 feature_fraction_stage2, val_score: 0.657825: 33%|###3 | 2/6 [00:01<00:03, 1.16it/s][I 2020-09-27 04:48:33,305] Trial 38 finished with value: 0.6578254229578112 and parameters: {'feature_fraction': 0.8160000000000001}. Best is trial 38 with value: 0.6578254229578112. feature_fraction_stage2, val_score: 0.657825: 33%|###3 | 2/6 [00:01<00:03, 1.16it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000906 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.649964 valid's binary_logloss: 0.664299 [200] train's binary_logloss: 0.638952 valid's binary_logloss: 0.66021 [300] train's binary_logloss: 0.632195 valid's binary_logloss: 0.6599 Early stopping, best iteration is: [265] train's binary_logloss: 0.634377 valid's binary_logloss: 0.659183 feature_fraction_stage2, val_score: 0.657825: 50%|##### | 3/6 [00:02<00:02, 1.11it/s][I 2020-09-27 04:48:34,293] Trial 39 finished with value: 0.6591825632130718 and parameters: {'feature_fraction': 0.88}. Best is trial 38 with value: 0.6578254229578112. feature_fraction_stage2, val_score: 0.657825: 50%|##### | 3/6 [00:02<00:02, 1.11it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016186 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650087 valid's binary_logloss: 0.664287 [200] train's binary_logloss: 0.639299 valid's binary_logloss: 0.66052 [300] train's binary_logloss: 0.632483 valid's binary_logloss: 0.659821 Early stopping, best iteration is: [265] train's binary_logloss: 0.634723 valid's binary_logloss: 0.659476 feature_fraction_stage2, val_score: 0.657825: 67%|######6 | 4/6 [00:03<00:01, 1.12it/s][I 2020-09-27 04:48:35,160] Trial 40 finished with value: 0.6594764202643758 and parameters: {'feature_fraction': 0.784}. Best is trial 38 with value: 0.6578254229578112. feature_fraction_stage2, val_score: 0.657825: 67%|######6 | 4/6 [00:03<00:01, 1.12it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005048 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650002 valid's binary_logloss: 0.663921 [200] train's binary_logloss: 0.639037 valid's binary_logloss: 0.659453 [300] train's binary_logloss: 0.632258 valid's binary_logloss: 0.658764 Early stopping, best iteration is: [263] train's binary_logloss: 0.634669 valid's binary_logloss: 0.658316 feature_fraction_stage2, val_score: 0.657825: 83%|########3 | 5/6 [00:04<00:00, 1.21it/s][I 2020-09-27 04:48:35,831] Trial 41 finished with value: 0.6583158730564244 and parameters: {'feature_fraction': 0.7200000000000001}. Best is trial 38 with value: 0.6578254229578112. feature_fraction_stage2, val_score: 0.657825: 83%|########3 | 5/6 [00:04<00:00, 1.21it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004831 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650087 valid's binary_logloss: 0.664287 [200] train's binary_logloss: 0.639299 valid's binary_logloss: 0.66052 [300] train's binary_logloss: 0.632483 valid's binary_logloss: 0.659821 Early stopping, best iteration is: [265] train's binary_logloss: 0.634723 valid's binary_logloss: 0.659476 feature_fraction_stage2, val_score: 0.657825: 100%|##########| 6/6 [00:04<00:00, 1.29it/s][I 2020-09-27 04:48:36,497] Trial 42 finished with value: 0.6594764202643758 and parameters: {'feature_fraction': 0.7520000000000001}. Best is trial 38 with value: 0.6578254229578112. feature_fraction_stage2, val_score: 0.657825: 100%|##########| 6/6 [00:04<00:00, 1.24it/s] regularization_factors, val_score: 0.657825: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000810 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650995 valid's binary_logloss: 0.663984 [200] train's binary_logloss: 0.640816 valid's binary_logloss: 0.660341 [300] train's binary_logloss: 0.635324 valid's binary_logloss: 0.659307 Early stopping, best iteration is: [279] train's binary_logloss: 0.636364 valid's binary_logloss: 0.659099 regularization_factors, val_score: 0.657825: 5%|5 | 1/20 [00:00<00:15, 1.24it/s][I 2020-09-27 04:48:37,323] Trial 43 finished with value: 0.659098571564232 and parameters: {'lambda_l1': 6.368584839611204, 'lambda_l2': 0.011340463712795283}. Best is trial 43 with value: 0.659098571564232. regularization_factors, val_score: 0.657825: 5%|5 | 1/20 [00:00<00:15, 1.24it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004832 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448 [200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404 [300] train's binary_logloss: 0.631873 valid's binary_logloss: 0.657967 Early stopping, best iteration is: [281] train's binary_logloss: 0.633131 valid's binary_logloss: 0.657825 regularization_factors, val_score: 0.657825: 10%|# | 2/20 [00:01<00:15, 1.13it/s][I 2020-09-27 04:48:38,394] Trial 44 finished with value: 0.6578254666437432 and parameters: {'lambda_l1': 2.7238144783184155e-08, 'lambda_l2': 1.008805152305263e-07}. Best is trial 44 with value: 0.6578254666437432. regularization_factors, val_score: 0.657825: 10%|# | 2/20 [00:01<00:15, 1.13it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.021204 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448 [200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404 [300] train's binary_logloss: 0.631873 valid's binary_logloss: 0.657967 Early stopping, best iteration is: [281] train's binary_logloss: 0.633131 valid's binary_logloss: 0.657825 regularization_factors, val_score: 0.657825: 15%|#5 | 3/20 [00:02<00:14, 1.13it/s][I 2020-09-27 04:48:39,260] Trial 45 finished with value: 0.6578254666410345 and parameters: {'lambda_l1': 1.0435559181670616e-08, 'lambda_l2': 1.585450495098862e-08}. Best is trial 45 with value: 0.6578254666410345. regularization_factors, val_score: 0.657825: 15%|#5 | 3/20 [00:02<00:14, 1.13it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000902 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448 [200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404 [300] train's binary_logloss: 0.631876 valid's binary_logloss: 0.657967 Early stopping, best iteration is: [281] train's binary_logloss: 0.633134 valid's binary_logloss: 0.657826 regularization_factors, val_score: 0.657825: 20%|## | 4/20 [00:03<00:13, 1.19it/s][I 2020-09-27 04:48:39,997] Trial 46 finished with value: 0.6578256353646856 and parameters: {'lambda_l1': 1.0219085602778263e-08, 'lambda_l2': 1.1221944608727081e-08}. Best is trial 45 with value: 0.6578254666410345. regularization_factors, val_score: 0.657825: 20%|## | 4/20 [00:03<00:13, 1.19it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001003 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448 [200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404 [300] train's binary_logloss: 0.631876 valid's binary_logloss: 0.657967 Early stopping, best iteration is: [281] train's binary_logloss: 0.633134 valid's binary_logloss: 0.657826 regularization_factors, val_score: 0.657825: 25%|##5 | 5/20 [00:04<00:12, 1.22it/s][I 2020-09-27 04:48:40,774] Trial 47 finished with value: 0.6578256353644696 and parameters: {'lambda_l1': 1.0248936810635019e-08, 'lambda_l2': 1.3858509908150066e-08}. Best is trial 45 with value: 0.6578254666410345. regularization_factors, val_score: 0.657825: 25%|##5 | 5/20 [00:04<00:12, 1.22it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000912 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448 [200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404 [300] train's binary_logloss: 0.631876 valid's binary_logloss: 0.657967 Early stopping, best iteration is: [281] train's binary_logloss: 0.633134 valid's binary_logloss: 0.657826 regularization_factors, val_score: 0.657825: 30%|### | 6/20 [00:04<00:11, 1.26it/s][I 2020-09-27 04:48:41,504] Trial 48 finished with value: 0.6578256353646091 and parameters: {'lambda_l1': 1.4330977464180962e-08, 'lambda_l2': 1.3975871503118737e-08}. Best is trial 45 with value: 0.6578254666410345. regularization_factors, val_score: 0.657825: 30%|### | 6/20 [00:04<00:11, 1.26it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008196 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448 [200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404 [300] train's binary_logloss: 0.631873 valid's binary_logloss: 0.657967 Early stopping, best iteration is: [281] train's binary_logloss: 0.633131 valid's binary_logloss: 0.657825 regularization_factors, val_score: 0.657825: 35%|###5 | 7/20 [00:06<00:11, 1.10it/s][I 2020-09-27 04:48:42,673] Trial 49 finished with value: 0.6578254666417223 and parameters: {'lambda_l1': 1.3796608447685168e-08, 'lambda_l2': 4.7462022137544425e-08}. Best is trial 45 with value: 0.6578254666410345. regularization_factors, val_score: 0.657825: 35%|###5 | 7/20 [00:06<00:11, 1.10it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000848 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448 [200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404 [300] train's binary_logloss: 0.631873 valid's binary_logloss: 0.657967 Early stopping, best iteration is: [281] train's binary_logloss: 0.633131 valid's binary_logloss: 0.657825 regularization_factors, val_score: 0.657825: 40%|#### | 8/20 [00:07<00:10, 1.13it/s][I 2020-09-27 04:48:43,519] Trial 50 finished with value: 0.6578254844386293 and parameters: {'lambda_l1': 5.179213600763938e-06, 'lambda_l2': 5.7935679948530245e-06}. Best is trial 45 with value: 0.6578254666410345. regularization_factors, val_score: 0.657825: 40%|#### | 8/20 [00:07<00:10, 1.13it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000955 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448 [200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404 [300] train's binary_logloss: 0.631876 valid's binary_logloss: 0.657967 Early stopping, best iteration is: [281] train's binary_logloss: 0.633134 valid's binary_logloss: 0.657825 regularization_factors, val_score: 0.657825: 45%|####5 | 9/20 [00:07<00:09, 1.19it/s][I 2020-09-27 04:48:44,255] Trial 51 finished with value: 0.6578252553356203 and parameters: {'lambda_l1': 1.558360532559241e-05, 'lambda_l2': 4.619266691410038e-06}. Best is trial 51 with value: 0.6578252553356203. regularization_factors, val_score: 0.657825: 45%|####5 | 9/20 [00:07<00:09, 1.19it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000894 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448 [200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404 [300] train's binary_logloss: 0.631876 valid's binary_logloss: 0.657967 Early stopping, best iteration is: [281] train's binary_logloss: 0.633134 valid's binary_logloss: 0.657826 regularization_factors, val_score: 0.657825: 50%|##### | 10/20 [00:08<00:08, 1.23it/s][I 2020-09-27 04:48:44,994] Trial 52 finished with value: 0.6578256356080326 and parameters: {'lambda_l1': 2.7250062441770353e-06, 'lambda_l2': 2.367717105214214e-06}. Best is trial 51 with value: 0.6578252553356203. regularization_factors, val_score: 0.657825: 50%|##### | 10/20 [00:08<00:08, 1.23it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001566 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448 [200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404 [300] train's binary_logloss: 0.631876 valid's binary_logloss: 0.657967 Early stopping, best iteration is: [281] train's binary_logloss: 0.633134 valid's binary_logloss: 0.657826 regularization_factors, val_score: 0.657825: 55%|#####5 | 11/20 [00:09<00:07, 1.26it/s][I 2020-09-27 04:48:45,738] Trial 53 finished with value: 0.6578256354185161 and parameters: {'lambda_l1': 3.3319990687032807e-07, 'lambda_l2': 9.864870800228018e-07}. Best is trial 51 with value: 0.6578252553356203. regularization_factors, val_score: 0.657825: 55%|#####5 | 11/20 [00:09<00:07, 1.26it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000843 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650018 valid's binary_logloss: 0.663448 [200] train's binary_logloss: 0.638999 valid's binary_logloss: 0.659122 [300] train's binary_logloss: 0.632026 valid's binary_logloss: 0.658235 Early stopping, best iteration is: [290] train's binary_logloss: 0.632696 valid's binary_logloss: 0.658147 regularization_factors, val_score: 0.657825: 60%|###### | 12/20 [00:10<00:07, 1.08it/s][I 2020-09-27 04:48:46,984] Trial 54 finished with value: 0.6581469442698293 and parameters: {'lambda_l1': 0.006687572450441869, 'lambda_l2': 2.42706687165855e-07}. Best is trial 51 with value: 0.6578252553356203. regularization_factors, val_score: 0.657825: 60%|###### | 12/20 [00:10<00:07, 1.08it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000800 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448 [200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404 [300] train's binary_logloss: 0.631879 valid's binary_logloss: 0.657967 Early stopping, best iteration is: [281] train's binary_logloss: 0.633137 valid's binary_logloss: 0.657825 regularization_factors, val_score: 0.657825: 65%|######5 | 13/20 [00:11<00:06, 1.14it/s][I 2020-09-27 04:48:47,745] Trial 55 finished with value: 0.6578254639325656 and parameters: {'lambda_l1': 0.0006287455216203462, 'lambda_l2': 8.873368461516288e-05}. Best is trial 51 with value: 0.6578252553356203. regularization_factors, val_score: 0.657825: 65%|######5 | 13/20 [00:11<00:06, 1.14it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000955 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448 [200] train's binary_logloss: 0.638998 valid's binary_logloss: 0.659122 [300] train's binary_logloss: 0.632035 valid's binary_logloss: 0.658192 [400] train's binary_logloss: 0.626308 valid's binary_logloss: 0.65838 Early stopping, best iteration is: [307] train's binary_logloss: 0.631536 valid's binary_logloss: 0.658173 regularization_factors, val_score: 0.657825: 70%|####### | 14/20 [00:12<00:05, 1.18it/s][I 2020-09-27 04:48:48,519] Trial 56 finished with value: 0.6581734662791846 and parameters: {'lambda_l1': 0.0023930535028393245, 'lambda_l2': 0.00017393958564666502}. Best is trial 51 with value: 0.6578252553356203. regularization_factors, val_score: 0.657825: 70%|####### | 14/20 [00:12<00:05, 1.18it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000968 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448 [200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404 [300] train's binary_logloss: 0.631879 valid's binary_logloss: 0.657967 Early stopping, best iteration is: [281] train's binary_logloss: 0.633137 valid's binary_logloss: 0.657825 regularization_factors, val_score: 0.657825: 75%|#######5 | 15/20 [00:12<00:04, 1.23it/s][I 2020-09-27 04:48:49,262] Trial 57 finished with value: 0.6578254362526145 and parameters: {'lambda_l1': 0.0001139415321377858, 'lambda_l2': 0.00018316916066158292}. Best is trial 51 with value: 0.6578252553356203. regularization_factors, val_score: 0.657825: 75%|#######5 | 15/20 [00:12<00:04, 1.23it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001132 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448 [200] train's binary_logloss: 0.63892 valid's binary_logloss: 0.659404 [300] train's binary_logloss: 0.63188 valid's binary_logloss: 0.657967 Early stopping, best iteration is: [281] train's binary_logloss: 0.633137 valid's binary_logloss: 0.657825 regularization_factors, val_score: 0.657825: 80%|######## | 16/20 [00:13<00:03, 1.16it/s][I 2020-09-27 04:48:50,236] Trial 58 finished with value: 0.6578254556273145 and parameters: {'lambda_l1': 0.00013528915615576415, 'lambda_l2': 0.0002063652161436333}. Best is trial 51 with value: 0.6578252553356203. regularization_factors, val_score: 0.657825: 80%|######## | 16/20 [00:13<00:03, 1.16it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.015940 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448 [200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404 [300] train's binary_logloss: 0.631876 valid's binary_logloss: 0.657967 Early stopping, best iteration is: [281] train's binary_logloss: 0.633134 valid's binary_logloss: 0.657826 regularization_factors, val_score: 0.657825: 85%|########5 | 17/20 [00:14<00:02, 1.13it/s][I 2020-09-27 04:48:51,159] Trial 59 finished with value: 0.6578256499831218 and parameters: {'lambda_l1': 0.00013602675042342853, 'lambda_l2': 0.00018327214045940748}. Best is trial 51 with value: 0.6578252553356203. regularization_factors, val_score: 0.657825: 85%|########5 | 17/20 [00:14<00:02, 1.13it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001556 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448 [200] train's binary_logloss: 0.63892 valid's binary_logloss: 0.659405 [300] train's binary_logloss: 0.631879 valid's binary_logloss: 0.657967 Early stopping, best iteration is: [281] train's binary_logloss: 0.633136 valid's binary_logloss: 0.657826 regularization_factors, val_score: 0.657825: 90%|######### | 18/20 [00:15<00:01, 1.20it/s][I 2020-09-27 04:48:51,873] Trial 60 finished with value: 0.6578258958216437 and parameters: {'lambda_l1': 0.00015253198331570547, 'lambda_l2': 0.0071388436435203435}. Best is trial 51 with value: 0.6578252553356203. regularization_factors, val_score: 0.657825: 90%|######### | 18/20 [00:15<00:01, 1.20it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000941 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448 [200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404 [300] train's binary_logloss: 0.63188 valid's binary_logloss: 0.657967 Early stopping, best iteration is: [281] train's binary_logloss: 0.633137 valid's binary_logloss: 0.657826 regularization_factors, val_score: 0.657825: 95%|#########5| 19/20 [00:16<00:00, 1.23it/s][I 2020-09-27 04:48:52,641] Trial 61 finished with value: 0.657825517008465 and parameters: {'lambda_l1': 0.0015382833109009928, 'lambda_l2': 4.130282202202608e-05}. Best is trial 51 with value: 0.6578252553356203. regularization_factors, val_score: 0.657825: 95%|#########5| 19/20 [00:16<00:00, 1.23it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000998 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448 [200] train's binary_logloss: 0.638919 valid's binary_logloss: 0.659404 [300] train's binary_logloss: 0.631874 valid's binary_logloss: 0.657967 Early stopping, best iteration is: [281] train's binary_logloss: 0.633131 valid's binary_logloss: 0.657826 regularization_factors, val_score: 0.657825: 100%|##########| 20/20 [00:16<00:00, 1.24it/s][I 2020-09-27 04:48:53,434] Trial 62 finished with value: 0.6578255316540529 and parameters: {'lambda_l1': 3.339088375221171e-05, 'lambda_l2': 0.0017874187322570777}. Best is trial 51 with value: 0.6578252553356203. regularization_factors, val_score: 0.657825: 100%|##########| 20/20 [00:16<00:00, 1.18it/s] min_data_in_leaf, val_score: 0.657825: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000868 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448 [200] train's binary_logloss: 0.638922 valid's binary_logloss: 0.659362 [300] train's binary_logloss: 0.632151 valid's binary_logloss: 0.658294 [400] train's binary_logloss: 0.626333 valid's binary_logloss: 0.659009 Early stopping, best iteration is: [300] train's binary_logloss: 0.632151 valid's binary_logloss: 0.658294 min_data_in_leaf, val_score: 0.657825: 20%|## | 1/5 [00:01<00:05, 1.28s/it][I 2020-09-27 04:48:54,726] Trial 63 finished with value: 0.6582938352043944 and parameters: {'min_child_samples': 25}. Best is trial 63 with value: 0.6582938352043944. min_data_in_leaf, val_score: 0.657825: 20%|## | 1/5 [00:01<00:05, 1.28s/it][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.005115 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448 [200] train's binary_logloss: 0.638826 valid's binary_logloss: 0.659409 [300] train's binary_logloss: 0.63179 valid's binary_logloss: 0.658711 Early stopping, best iteration is: [232] train's binary_logloss: 0.636374 valid's binary_logloss: 0.658604 min_data_in_leaf, val_score: 0.657825: 40%|#### | 2/5 [00:01<00:03, 1.10s/it][I 2020-09-27 04:48:55,420] Trial 64 finished with value: 0.6586044923878769 and parameters: {'min_child_samples': 5}. Best is trial 63 with value: 0.6582938352043944. min_data_in_leaf, val_score: 0.657825: 40%|#### | 2/5 [00:01<00:03, 1.10s/it][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000876 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.65002 valid's binary_logloss: 0.663404 [200] train's binary_logloss: 0.639166 valid's binary_logloss: 0.659615 [300] train's binary_logloss: 0.632623 valid's binary_logloss: 0.659448 Early stopping, best iteration is: [240] train's binary_logloss: 0.636379 valid's binary_logloss: 0.65892 min_data_in_leaf, val_score: 0.657825: 60%|###### | 3/5 [00:02<00:01, 1.02it/s][I 2020-09-27 04:48:56,123] Trial 65 finished with value: 0.658919688761251 and parameters: {'min_child_samples': 50}. Best is trial 63 with value: 0.6582938352043944. min_data_in_leaf, val_score: 0.657825: 60%|###### | 3/5 [00:02<00:01, 1.02it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.004999 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.65002 valid's binary_logloss: 0.663404 [200] train's binary_logloss: 0.639125 valid's binary_logloss: 0.659535 [300] train's binary_logloss: 0.632949 valid's binary_logloss: 0.658458 Early stopping, best iteration is: [266] train's binary_logloss: 0.634844 valid's binary_logloss: 0.658133 min_data_in_leaf, val_score: 0.657825: 80%|######## | 4/5 [00:03<00:00, 1.10it/s][I 2020-09-27 04:48:56,850] Trial 66 finished with value: 0.6581334712449208 and parameters: {'min_child_samples': 100}. Best is trial 66 with value: 0.6581334712449208. min_data_in_leaf, val_score: 0.657825: 80%|######## | 4/5 [00:03<00:00, 1.10it/s][LightGBM] [Info] Number of positive: 12844, number of negative: 13156 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000869 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4241 [LightGBM] [Info] Number of data points in the train set: 26000, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.494000 -> initscore=-0.024001 [LightGBM] [Info] Start training from score -0.024001 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.650017 valid's binary_logloss: 0.663448 [200] train's binary_logloss: 0.638826 valid's binary_logloss: 0.659409 [300] train's binary_logloss: 0.631757 valid's binary_logloss: 0.658712 Early stopping, best iteration is: [234] train's binary_logloss: 0.636229 valid's binary_logloss: 0.65847 min_data_in_leaf, val_score: 0.657825: 100%|##########| 5/5 [00:04<00:00, 1.19it/s][I 2020-09-27 04:48:57,537] Trial 67 finished with value: 0.6584702414733883 and parameters: {'min_child_samples': 10}. Best is trial 66 with value: 0.6581334712449208. min_data_in_leaf, val_score: 0.657825: 100%|##########| 5/5 [00:04<00:00, 1.22it/s] ################################ CV_score:0.6150646280811062 Fold : 0 [I 2020-09-27 04:48:57,616] A new study created in memory with name: no-name-8ff89ddf-7641-453c-86d0-fc507cfacb11 feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.026169 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663649 valid's binary_logloss: 0.690213 Early stopping, best iteration is: [45] train's binary_logloss: 0.676671 valid's binary_logloss: 0.689632 feature_fraction, val_score: 0.689632: 14%|#4 | 1/7 [00:01<00:08, 1.45s/it][I 2020-09-27 04:48:59,075] Trial 0 finished with value: 0.6896322280348909 and parameters: {'feature_fraction': 0.7}. Best is trial 0 with value: 0.6896322280348909. feature_fraction, val_score: 0.689632: 14%|#4 | 1/7 [00:01<00:08, 1.45s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001410 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.66617 valid's binary_logloss: 0.690438 Early stopping, best iteration is: [41] train's binary_logloss: 0.679097 valid's binary_logloss: 0.689783 feature_fraction, val_score: 0.689632: 29%|##8 | 2/7 [00:02<00:06, 1.24s/it][I 2020-09-27 04:48:59,837] Trial 1 finished with value: 0.6897826459094011 and parameters: {'feature_fraction': 0.4}. Best is trial 0 with value: 0.6896322280348909. feature_fraction, val_score: 0.689632: 29%|##8 | 2/7 [00:02<00:06, 1.24s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.003154 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.662081 valid's binary_logloss: 0.690003 Early stopping, best iteration is: [41] train's binary_logloss: 0.677046 valid's binary_logloss: 0.68968 feature_fraction, val_score: 0.689632: 43%|####2 | 3/7 [00:03<00:04, 1.13s/it][I 2020-09-27 04:49:00,694] Trial 2 finished with value: 0.6896804566917485 and parameters: {'feature_fraction': 1.0}. Best is trial 0 with value: 0.6896322280348909. feature_fraction, val_score: 0.689632: 43%|####2 | 3/7 [00:03<00:04, 1.13s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.015461 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.66415 valid's binary_logloss: 0.689593 Early stopping, best iteration is: [39] train's binary_logloss: 0.678462 valid's binary_logloss: 0.689021 feature_fraction, val_score: 0.689021: 57%|#####7 | 4/7 [00:03<00:03, 1.01s/it][I 2020-09-27 04:49:01,437] Trial 3 finished with value: 0.6890207233896298 and parameters: {'feature_fraction': 0.6}. Best is trial 3 with value: 0.6890207233896298. feature_fraction, val_score: 0.689021: 57%|#####7 | 4/7 [00:03<00:03, 1.01s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.025191 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.66251 valid's binary_logloss: 0.689598 Early stopping, best iteration is: [42] train's binary_logloss: 0.676825 valid's binary_logloss: 0.68942 feature_fraction, val_score: 0.689021: 71%|#######1 | 5/7 [00:05<00:02, 1.13s/it][I 2020-09-27 04:49:02,837] Trial 4 finished with value: 0.6894197474355058 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 3 with value: 0.6890207233896298. feature_fraction, val_score: 0.689021: 71%|#######1 | 5/7 [00:05<00:02, 1.13s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.017648 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663408 valid's binary_logloss: 0.690473 Early stopping, best iteration is: [45] train's binary_logloss: 0.676529 valid's binary_logloss: 0.689828 feature_fraction, val_score: 0.689021: 86%|########5 | 6/7 [00:06<00:01, 1.04s/it][I 2020-09-27 04:49:03,685] Trial 5 finished with value: 0.68982789311769 and parameters: {'feature_fraction': 0.8}. Best is trial 3 with value: 0.6890207233896298. feature_fraction, val_score: 0.689021: 86%|########5 | 6/7 [00:06<00:01, 1.04s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001730 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665183 valid's binary_logloss: 0.689631 [200] train's binary_logloss: 0.645808 valid's binary_logloss: 0.690591 Early stopping, best iteration is: [102] train's binary_logloss: 0.664779 valid's binary_logloss: 0.689546 feature_fraction, val_score: 0.689021: 100%|##########| 7/7 [00:07<00:00, 1.07s/it][I 2020-09-27 04:49:04,813] Trial 6 finished with value: 0.6895458054912846 and parameters: {'feature_fraction': 0.5}. Best is trial 3 with value: 0.6890207233896298. feature_fraction, val_score: 0.689021: 100%|##########| 7/7 [00:07<00:00, 1.03s/it] num_leaves, val_score: 0.689021: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012666 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.577872 valid's binary_logloss: 0.693632 Early stopping, best iteration is: [18] train's binary_logloss: 0.663683 valid's binary_logloss: 0.689631 num_leaves, val_score: 0.689021: 5%|5 | 1/20 [00:01<00:30, 1.61s/it][I 2020-09-27 04:49:06,444] Trial 7 finished with value: 0.689631097968691 and parameters: {'num_leaves': 168}. Best is trial 7 with value: 0.689631097968691. num_leaves, val_score: 0.689021: 5%|5 | 1/20 [00:01<00:30, 1.61s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.017419 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.604202 valid's binary_logloss: 0.69167 Early stopping, best iteration is: [38] train's binary_logloss: 0.650476 valid's binary_logloss: 0.689618 num_leaves, val_score: 0.689021: 10%|# | 2/20 [00:02<00:26, 1.47s/it][I 2020-09-27 04:49:07,566] Trial 8 finished with value: 0.689618324134334 and parameters: {'num_leaves': 122}. Best is trial 8 with value: 0.689618324134334. num_leaves, val_score: 0.689021: 10%|# | 2/20 [00:02<00:26, 1.47s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016515 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685172 valid's binary_logloss: 0.689666 [200] train's binary_logloss: 0.681255 valid's binary_logloss: 0.689678 Early stopping, best iteration is: [156] train's binary_logloss: 0.68289 valid's binary_logloss: 0.689534 num_leaves, val_score: 0.689021: 15%|#5 | 3/20 [00:03<00:22, 1.34s/it][I 2020-09-27 04:49:08,626] Trial 9 finished with value: 0.6895342400114649 and parameters: {'num_leaves': 6}. Best is trial 9 with value: 0.6895342400114649. num_leaves, val_score: 0.689021: 15%|#5 | 3/20 [00:03<00:22, 1.34s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016558 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.535492 valid's binary_logloss: 0.695541 Early stopping, best iteration is: [19] train's binary_logloss: 0.649759 valid's binary_logloss: 0.690794 num_leaves, val_score: 0.689021: 20%|## | 4/20 [00:05<00:23, 1.49s/it][I 2020-09-27 04:49:10,461] Trial 10 finished with value: 0.6907940845828584 and parameters: {'num_leaves': 253}. Best is trial 9 with value: 0.6895342400114649. num_leaves, val_score: 0.689021: 20%|## | 4/20 [00:05<00:23, 1.49s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.017448 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.67101 valid's binary_logloss: 0.689552 Early stopping, best iteration is: [93] train's binary_logloss: 0.672156 valid's binary_logloss: 0.68942 num_leaves, val_score: 0.689021: 25%|##5 | 5/20 [00:06<00:19, 1.31s/it][I 2020-09-27 04:49:11,362] Trial 11 finished with value: 0.6894197097930296 and parameters: {'num_leaves': 22}. Best is trial 11 with value: 0.6894197097930296. num_leaves, val_score: 0.689021: 25%|##5 | 5/20 [00:06<00:19, 1.31s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.014207 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.690044 valid's binary_logloss: 0.690612 [200] train's binary_logloss: 0.68885 valid's binary_logloss: 0.689962 [300] train's binary_logloss: 0.688098 valid's binary_logloss: 0.689612 [400] train's binary_logloss: 0.687569 valid's binary_logloss: 0.689487 [500] train's binary_logloss: 0.68718 valid's binary_logloss: 0.689391 [600] train's binary_logloss: 0.686881 valid's binary_logloss: 0.689318 [700] train's binary_logloss: 0.68664 valid's binary_logloss: 0.689265 [800] train's binary_logloss: 0.68644 valid's binary_logloss: 0.689264 Early stopping, best iteration is: [710] train's binary_logloss: 0.686618 valid's binary_logloss: 0.689244 num_leaves, val_score: 0.689021: 30%|### | 6/20 [00:09<00:25, 1.86s/it][I 2020-09-27 04:49:14,484] Trial 12 finished with value: 0.689244475322615 and parameters: {'num_leaves': 2}. Best is trial 12 with value: 0.689244475322615. num_leaves, val_score: 0.689021: 30%|### | 6/20 [00:09<00:25, 1.86s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.015182 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.630223 valid's binary_logloss: 0.692152 Early stopping, best iteration is: [21] train's binary_logloss: 0.674331 valid's binary_logloss: 0.690049 num_leaves, val_score: 0.689021: 35%|###5 | 7/20 [00:10<00:20, 1.56s/it][I 2020-09-27 04:49:15,354] Trial 13 finished with value: 0.6900491534448284 and parameters: {'num_leaves': 80}. Best is trial 12 with value: 0.689244475322615. num_leaves, val_score: 0.689021: 35%|###5 | 7/20 [00:10<00:20, 1.56s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.017787 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.543727 valid's binary_logloss: 0.696815 Early stopping, best iteration is: [10] train's binary_logloss: 0.670334 valid's binary_logloss: 0.691482 num_leaves, val_score: 0.689021: 40%|#### | 8/20 [00:11<00:17, 1.46s/it][I 2020-09-27 04:49:16,588] Trial 14 finished with value: 0.6914823998645343 and parameters: {'num_leaves': 234}. Best is trial 12 with value: 0.689244475322615. num_leaves, val_score: 0.689021: 40%|#### | 8/20 [00:11<00:17, 1.46s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.015457 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.646446 valid's binary_logloss: 0.691884 Early stopping, best iteration is: [17] train's binary_logloss: 0.681433 valid's binary_logloss: 0.689734 num_leaves, val_score: 0.689021: 45%|####5 | 9/20 [00:13<00:15, 1.43s/it][I 2020-09-27 04:49:17,939] Trial 15 finished with value: 0.6897344485270362 and parameters: {'num_leaves': 55}. Best is trial 12 with value: 0.689244475322615. num_leaves, val_score: 0.689021: 45%|####5 | 9/20 [00:13<00:15, 1.43s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.015412 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.570477 valid's binary_logloss: 0.695696 Early stopping, best iteration is: [23] train's binary_logloss: 0.654169 valid's binary_logloss: 0.690909 num_leaves, val_score: 0.689021: 50%|##### | 10/20 [00:14<00:13, 1.34s/it][I 2020-09-27 04:49:19,067] Trial 16 finished with value: 0.6909092950799796 and parameters: {'num_leaves': 181}. Best is trial 12 with value: 0.689244475322615. num_leaves, val_score: 0.689021: 50%|##### | 10/20 [00:14<00:13, 1.34s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016435 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.623443 valid's binary_logloss: 0.691768 Early stopping, best iteration is: [19] train's binary_logloss: 0.674343 valid's binary_logloss: 0.690276 num_leaves, val_score: 0.689021: 55%|#####5 | 11/20 [00:15<00:10, 1.20s/it][I 2020-09-27 04:49:19,941] Trial 17 finished with value: 0.690276336147683 and parameters: {'num_leaves': 90}. Best is trial 12 with value: 0.689244475322615. num_leaves, val_score: 0.689021: 55%|#####5 | 11/20 [00:15<00:10, 1.20s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016181 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.659158 valid's binary_logloss: 0.689655 Early stopping, best iteration is: [35] train's binary_logloss: 0.677405 valid's binary_logloss: 0.689049 num_leaves, val_score: 0.689021: 60%|###### | 12/20 [00:15<00:08, 1.07s/it][I 2020-09-27 04:49:20,717] Trial 18 finished with value: 0.6890491443530252 and parameters: {'num_leaves': 38}. Best is trial 18 with value: 0.6890491443530252. num_leaves, val_score: 0.689021: 60%|###### | 12/20 [00:15<00:08, 1.07s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.023330 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.652371 valid's binary_logloss: 0.690333 Early stopping, best iteration is: [39] train's binary_logloss: 0.672949 valid's binary_logloss: 0.68919 num_leaves, val_score: 0.689021: 65%|######5 | 13/20 [00:17<00:08, 1.17s/it][I 2020-09-27 04:49:22,101] Trial 19 finished with value: 0.6891896975164848 and parameters: {'num_leaves': 47}. Best is trial 18 with value: 0.6890491443530252. num_leaves, val_score: 0.689021: 65%|######5 | 13/20 [00:17<00:08, 1.17s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.017299 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.59964 valid's binary_logloss: 0.69374 Early stopping, best iteration is: [14] train's binary_logloss: 0.674012 valid's binary_logloss: 0.690104 num_leaves, val_score: 0.689021: 70%|####### | 14/20 [00:18<00:06, 1.10s/it][I 2020-09-27 04:49:23,062] Trial 20 finished with value: 0.6901038597516463 and parameters: {'num_leaves': 130}. Best is trial 18 with value: 0.6890491443530252. num_leaves, val_score: 0.689021: 70%|####### | 14/20 [00:18<00:06, 1.10s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016669 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657163 valid's binary_logloss: 0.690788 Early stopping, best iteration is: [43] train's binary_logloss: 0.674068 valid's binary_logloss: 0.689599 num_leaves, val_score: 0.689021: 75%|#######5 | 15/20 [00:19<00:05, 1.01s/it][I 2020-09-27 04:49:23,851] Trial 21 finished with value: 0.6895994157631041 and parameters: {'num_leaves': 40}. Best is trial 18 with value: 0.6890491443530252. num_leaves, val_score: 0.689021: 75%|#######5 | 15/20 [00:19<00:05, 1.01s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016387 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.632347 valid's binary_logloss: 0.692318 Early stopping, best iteration is: [23] train's binary_logloss: 0.673649 valid's binary_logloss: 0.689645 num_leaves, val_score: 0.689021: 80%|######## | 16/20 [00:19<00:03, 1.05it/s][I 2020-09-27 04:49:24,677] Trial 22 finished with value: 0.6896448463652016 and parameters: {'num_leaves': 76}. Best is trial 18 with value: 0.6890491443530252. num_leaves, val_score: 0.689021: 80%|######## | 16/20 [00:19<00:03, 1.05it/s][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.020197 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656137 valid's binary_logloss: 0.689636 Early stopping, best iteration is: [96] train's binary_logloss: 0.657272 valid's binary_logloss: 0.689437 num_leaves, val_score: 0.689021: 85%|########5 | 17/20 [00:21<00:03, 1.10s/it][I 2020-09-27 04:49:26,120] Trial 23 finished with value: 0.6894373899731695 and parameters: {'num_leaves': 42}. Best is trial 18 with value: 0.6890491443530252. num_leaves, val_score: 0.689021: 85%|########5 | 17/20 [00:21<00:03, 1.10s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016258 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.608119 valid's binary_logloss: 0.691911 Early stopping, best iteration is: [30] train's binary_logloss: 0.659223 valid's binary_logloss: 0.689435 num_leaves, val_score: 0.689021: 90%|######### | 18/20 [00:22<00:02, 1.09s/it][I 2020-09-27 04:49:27,187] Trial 24 finished with value: 0.6894354954462946 and parameters: {'num_leaves': 116}. Best is trial 18 with value: 0.6890491443530252. num_leaves, val_score: 0.689021: 90%|######### | 18/20 [00:22<00:02, 1.09s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.018058 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68615 valid's binary_logloss: 0.689604 [200] train's binary_logloss: 0.682919 valid's binary_logloss: 0.689378 Early stopping, best iteration is: [158] train's binary_logloss: 0.684169 valid's binary_logloss: 0.689219 num_leaves, val_score: 0.689021: 95%|#########5| 19/20 [00:23<00:01, 1.08s/it][I 2020-09-27 04:49:28,242] Trial 25 finished with value: 0.6892189395196165 and parameters: {'num_leaves': 5}. Best is trial 18 with value: 0.6890491443530252. num_leaves, val_score: 0.689021: 95%|#########5| 19/20 [00:23<00:01, 1.08s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016101 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.651145 valid's binary_logloss: 0.690351 Early stopping, best iteration is: [34] train's binary_logloss: 0.674412 valid's binary_logloss: 0.689378 num_leaves, val_score: 0.689021: 100%|##########| 20/20 [00:24<00:00, 1.16s/it][I 2020-09-27 04:49:29,591] Trial 26 finished with value: 0.6893783362619839 and parameters: {'num_leaves': 49}. Best is trial 18 with value: 0.6890491443530252. num_leaves, val_score: 0.689021: 100%|##########| 20/20 [00:24<00:00, 1.24s/it] bagging, val_score: 0.689021: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013092 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664277 valid's binary_logloss: 0.68939 Early stopping, best iteration is: [71] train's binary_logloss: 0.670651 valid's binary_logloss: 0.688825 bagging, val_score: 0.688825: 10%|# | 1/10 [00:01<00:09, 1.08s/it][I 2020-09-27 04:49:30,680] Trial 27 finished with value: 0.6888247862401992 and parameters: {'bagging_fraction': 0.958245225148382, 'bagging_freq': 2}. Best is trial 27 with value: 0.6888247862401992. bagging, val_score: 0.688825: 10%|# | 1/10 [00:01<00:09, 1.08s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.014041 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664031 valid's binary_logloss: 0.689087 [200] train's binary_logloss: 0.64388 valid's binary_logloss: 0.689768 Early stopping, best iteration is: [119] train's binary_logloss: 0.659997 valid's binary_logloss: 0.688879 bagging, val_score: 0.688825: 20%|## | 2/10 [00:02<00:09, 1.15s/it][I 2020-09-27 04:49:32,003] Trial 28 finished with value: 0.6888794888870088 and parameters: {'bagging_fraction': 0.9745012448452529, 'bagging_freq': 2}. Best is trial 27 with value: 0.6888247862401992. bagging, val_score: 0.688825: 20%|## | 2/10 [00:02<00:09, 1.15s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.015566 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664149 valid's binary_logloss: 0.689861 Early stopping, best iteration is: [50] train's binary_logloss: 0.675702 valid's binary_logloss: 0.689193 bagging, val_score: 0.688825: 30%|### | 3/10 [00:03<00:08, 1.23s/it][I 2020-09-27 04:49:33,428] Trial 29 finished with value: 0.689193182327793 and parameters: {'bagging_fraction': 0.9667819838074765, 'bagging_freq': 2}. Best is trial 27 with value: 0.6888247862401992. bagging, val_score: 0.688825: 30%|### | 3/10 [00:03<00:08, 1.23s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013341 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664296 valid's binary_logloss: 0.689547 Early stopping, best iteration is: [60] train's binary_logloss: 0.673228 valid's binary_logloss: 0.689074 bagging, val_score: 0.688825: 40%|#### | 4/10 [00:04<00:07, 1.17s/it][I 2020-09-27 04:49:34,451] Trial 30 finished with value: 0.6890740654309894 and parameters: {'bagging_fraction': 0.9690126597280045, 'bagging_freq': 2}. Best is trial 27 with value: 0.6888247862401992. bagging, val_score: 0.688825: 40%|#### | 4/10 [00:04<00:07, 1.17s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016944 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664297 valid's binary_logloss: 0.690412 Early stopping, best iteration is: [54] train's binary_logloss: 0.674733 valid's binary_logloss: 0.68958 bagging, val_score: 0.688825: 50%|##### | 5/10 [00:05<00:05, 1.10s/it][I 2020-09-27 04:49:35,398] Trial 31 finished with value: 0.6895802218638041 and parameters: {'bagging_fraction': 0.9943804516504096, 'bagging_freq': 4}. Best is trial 27 with value: 0.6888247862401992. bagging, val_score: 0.688825: 50%|##### | 5/10 [00:05<00:05, 1.10s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.014817 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664664 valid's binary_logloss: 0.689957 Early stopping, best iteration is: [37] train's binary_logloss: 0.679397 valid's binary_logloss: 0.689326 bagging, val_score: 0.688825: 60%|###### | 6/10 [00:06<00:04, 1.07s/it][I 2020-09-27 04:49:36,386] Trial 32 finished with value: 0.6893263473853064 and parameters: {'bagging_fraction': 0.6476701297352184, 'bagging_freq': 1}. Best is trial 27 with value: 0.6888247862401992. bagging, val_score: 0.688825: 60%|###### | 6/10 [00:06<00:04, 1.07s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.029663 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515 Early stopping, best iteration is: [49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719 bagging, val_score: 0.688719: 70%|####### | 7/10 [00:08<00:03, 1.16s/it][I 2020-09-27 04:49:37,766] Trial 33 finished with value: 0.6887188408517487 and parameters: {'bagging_fraction': 0.8287757616049252, 'bagging_freq': 7}. Best is trial 33 with value: 0.6887188408517487. bagging, val_score: 0.688719: 70%|####### | 7/10 [00:08<00:03, 1.16s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012644 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.66436 valid's binary_logloss: 0.689716 Early stopping, best iteration is: [51] train's binary_logloss: 0.675497 valid's binary_logloss: 0.689039 bagging, val_score: 0.688719: 80%|######## | 8/10 [00:09<00:02, 1.09s/it][I 2020-09-27 04:49:38,691] Trial 34 finished with value: 0.689039279019024 and parameters: {'bagging_fraction': 0.8282760453593162, 'bagging_freq': 7}. Best is trial 33 with value: 0.6887188408517487. bagging, val_score: 0.688719: 80%|######## | 8/10 [00:09<00:02, 1.09s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016881 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663999 valid's binary_logloss: 0.690093 Early stopping, best iteration is: [92] train's binary_logloss: 0.665835 valid's binary_logloss: 0.689992 bagging, val_score: 0.688719: 90%|######### | 9/10 [00:10<00:01, 1.10s/it][I 2020-09-27 04:49:39,803] Trial 35 finished with value: 0.6899923113836022 and parameters: {'bagging_fraction': 0.8214825248241318, 'bagging_freq': 4}. Best is trial 33 with value: 0.6887188408517487. bagging, val_score: 0.688719: 90%|######### | 9/10 [00:10<00:01, 1.10s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013803 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664239 valid's binary_logloss: 0.688911 [200] train's binary_logloss: 0.644408 valid's binary_logloss: 0.690305 Early stopping, best iteration is: [103] train's binary_logloss: 0.663554 valid's binary_logloss: 0.688799 bagging, val_score: 0.688719: 100%|##########| 10/10 [00:11<00:00, 1.24s/it][I 2020-09-27 04:49:41,362] Trial 36 finished with value: 0.6887991231168928 and parameters: {'bagging_fraction': 0.8383907363062589, 'bagging_freq': 7}. Best is trial 33 with value: 0.6887188408517487. bagging, val_score: 0.688719: 100%|##########| 10/10 [00:11<00:00, 1.18s/it] feature_fraction_stage2, val_score: 0.688719: 0%| | 0/6 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013665 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515 Early stopping, best iteration is: [49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719 feature_fraction_stage2, val_score: 0.688719: 17%|#6 | 1/6 [00:00<00:04, 1.03it/s][I 2020-09-27 04:49:42,349] Trial 37 finished with value: 0.6887188408517487 and parameters: {'feature_fraction': 0.616}. Best is trial 37 with value: 0.6887188408517487. feature_fraction_stage2, val_score: 0.688719: 17%|#6 | 1/6 [00:00<00:04, 1.03it/s][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001729 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664799 valid's binary_logloss: 0.689634 Early stopping, best iteration is: [69] train's binary_logloss: 0.671673 valid's binary_logloss: 0.689345 feature_fraction_stage2, val_score: 0.688719: 33%|###3 | 2/6 [00:02<00:03, 1.00it/s][I 2020-09-27 04:49:43,417] Trial 38 finished with value: 0.6893450614153805 and parameters: {'feature_fraction': 0.552}. Best is trial 37 with value: 0.6887188408517487. feature_fraction_stage2, val_score: 0.688719: 33%|###3 | 2/6 [00:02<00:03, 1.00it/s][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001868 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664799 valid's binary_logloss: 0.689634 Early stopping, best iteration is: [69] train's binary_logloss: 0.671673 valid's binary_logloss: 0.689345 feature_fraction_stage2, val_score: 0.688719: 50%|##### | 3/6 [00:03<00:03, 1.17s/it][I 2020-09-27 04:49:44,990] Trial 39 finished with value: 0.6893450614153805 and parameters: {'feature_fraction': 0.52}. Best is trial 37 with value: 0.6887188408517487. feature_fraction_stage2, val_score: 0.688719: 50%|##### | 3/6 [00:03<00:03, 1.17s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001814 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664285 valid's binary_logloss: 0.690298 Early stopping, best iteration is: [64] train's binary_logloss: 0.672325 valid's binary_logloss: 0.689397 feature_fraction_stage2, val_score: 0.688719: 67%|######6 | 4/6 [00:04<00:02, 1.14s/it][I 2020-09-27 04:49:46,067] Trial 40 finished with value: 0.6893965384793862 and parameters: {'feature_fraction': 0.584}. Best is trial 37 with value: 0.6887188408517487. feature_fraction_stage2, val_score: 0.688719: 67%|######6 | 4/6 [00:04<00:02, 1.14s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013983 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663794 valid's binary_logloss: 0.689901 [200] train's binary_logloss: 0.64382 valid's binary_logloss: 0.690825 Early stopping, best iteration is: [115] train's binary_logloss: 0.66065 valid's binary_logloss: 0.689797 feature_fraction_stage2, val_score: 0.688719: 83%|########3 | 5/6 [00:06<00:01, 1.42s/it][I 2020-09-27 04:49:48,122] Trial 41 finished with value: 0.6897973746145477 and parameters: {'feature_fraction': 0.6799999999999999}. Best is trial 37 with value: 0.6887188408517487. feature_fraction_stage2, val_score: 0.688719: 83%|########3 | 5/6 [00:06<00:01, 1.42s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.031387 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663746 valid's binary_logloss: 0.689073 [200] train's binary_logloss: 0.643757 valid's binary_logloss: 0.691217 Early stopping, best iteration is: [105] train's binary_logloss: 0.662622 valid's binary_logloss: 0.689054 feature_fraction_stage2, val_score: 0.688719: 100%|##########| 6/6 [00:08<00:00, 1.44s/it][I 2020-09-27 04:49:49,607] Trial 42 finished with value: 0.6890541064139101 and parameters: {'feature_fraction': 0.6479999999999999}. Best is trial 37 with value: 0.6887188408517487. feature_fraction_stage2, val_score: 0.688719: 100%|##########| 6/6 [00:08<00:00, 1.37s/it] regularization_factors, val_score: 0.688719: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012472 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664985 valid's binary_logloss: 0.690223 [200] train's binary_logloss: 0.646149 valid's binary_logloss: 0.690617 Early stopping, best iteration is: [151] train's binary_logloss: 0.655125 valid's binary_logloss: 0.690101 regularization_factors, val_score: 0.688719: 5%|5 | 1/20 [00:01<00:28, 1.50s/it][I 2020-09-27 04:49:51,126] Trial 43 finished with value: 0.6901011755065565 and parameters: {'lambda_l1': 1.1303470188424833, 'lambda_l2': 0.06375127708462985}. Best is trial 43 with value: 0.6901011755065565. regularization_factors, val_score: 0.688719: 5%|5 | 1/20 [00:01<00:28, 1.50s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.014406 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515 Early stopping, best iteration is: [49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719 regularization_factors, val_score: 0.688719: 10%|# | 2/20 [00:02<00:26, 1.47s/it][I 2020-09-27 04:49:52,538] Trial 44 finished with value: 0.6887188408527825 and parameters: {'lambda_l1': 1.8636898004930976e-08, 'lambda_l2': 6.192880162509011e-08}. Best is trial 44 with value: 0.6887188408527825. regularization_factors, val_score: 0.688719: 10%|# | 2/20 [00:02<00:26, 1.47s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016047 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515 Early stopping, best iteration is: [49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719 regularization_factors, val_score: 0.688719: 15%|#5 | 3/20 [00:03<00:22, 1.32s/it][I 2020-09-27 04:49:53,496] Trial 45 finished with value: 0.6887188408522336 and parameters: {'lambda_l1': 1.8439298966850396e-08, 'lambda_l2': 1.3511479384741456e-08}. Best is trial 45 with value: 0.6887188408522336. regularization_factors, val_score: 0.688719: 15%|#5 | 3/20 [00:03<00:22, 1.32s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012081 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515 Early stopping, best iteration is: [49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719 regularization_factors, val_score: 0.688719: 20%|## | 4/20 [00:04<00:19, 1.20s/it][I 2020-09-27 04:49:54,422] Trial 46 finished with value: 0.6887188408519217 and parameters: {'lambda_l1': 1.2530425311422802e-08, 'lambda_l2': 1.0756585467257088e-08}. Best is trial 46 with value: 0.6887188408519217. regularization_factors, val_score: 0.688719: 20%|## | 4/20 [00:04<00:19, 1.20s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.014442 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515 Early stopping, best iteration is: [49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719 regularization_factors, val_score: 0.688719: 25%|##5 | 5/20 [00:05<00:16, 1.13s/it][I 2020-09-27 04:49:55,370] Trial 47 finished with value: 0.6887188408519394 and parameters: {'lambda_l1': 1.2148997058183217e-08, 'lambda_l2': 1.076670603152709e-08}. Best is trial 46 with value: 0.6887188408519217. regularization_factors, val_score: 0.688719: 25%|##5 | 5/20 [00:05<00:16, 1.13s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.015325 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515 Early stopping, best iteration is: [49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719 regularization_factors, val_score: 0.688719: 30%|### | 6/20 [00:07<00:16, 1.21s/it][I 2020-09-27 04:49:56,764] Trial 48 finished with value: 0.6887188408519016 and parameters: {'lambda_l1': 1.0473302644937364e-08, 'lambda_l2': 1.4000255193861959e-08}. Best is trial 48 with value: 0.6887188408519016. regularization_factors, val_score: 0.688719: 30%|### | 6/20 [00:07<00:16, 1.21s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.017485 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515 Early stopping, best iteration is: [49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719 regularization_factors, val_score: 0.688719: 35%|###5 | 7/20 [00:08<00:14, 1.15s/it][I 2020-09-27 04:49:57,778] Trial 49 finished with value: 0.6887188408520548 and parameters: {'lambda_l1': 1.2933235345738932e-08, 'lambda_l2': 1.3210044579433778e-08}. Best is trial 48 with value: 0.6887188408519016. regularization_factors, val_score: 0.688719: 35%|###5 | 7/20 [00:08<00:14, 1.15s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016188 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515 Early stopping, best iteration is: [49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719 regularization_factors, val_score: 0.688719: 40%|#### | 8/20 [00:09<00:13, 1.09s/it][I 2020-09-27 04:49:58,722] Trial 50 finished with value: 0.6887188410169663 and parameters: {'lambda_l1': 6.522515649363731e-06, 'lambda_l2': 5.399603910938485e-06}. Best is trial 48 with value: 0.6887188408519016. regularization_factors, val_score: 0.688719: 40%|#### | 8/20 [00:09<00:13, 1.09s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.018393 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515 Early stopping, best iteration is: [49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719 regularization_factors, val_score: 0.688719: 45%|####5 | 9/20 [00:10<00:12, 1.18s/it][I 2020-09-27 04:50:00,120] Trial 51 finished with value: 0.6887188408522066 and parameters: {'lambda_l1': 1.7781053835321766e-08, 'lambda_l2': 1.2911207447672734e-08}. Best is trial 48 with value: 0.6887188408519016. regularization_factors, val_score: 0.688719: 45%|####5 | 9/20 [00:10<00:12, 1.18s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012528 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515 Early stopping, best iteration is: [49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719 regularization_factors, val_score: 0.688719: 50%|##### | 10/20 [00:11<00:11, 1.14s/it][I 2020-09-27 04:50:01,154] Trial 52 finished with value: 0.6887188408521625 and parameters: {'lambda_l1': 1.487838960944784e-08, 'lambda_l2': 1.2171741979090901e-08}. Best is trial 48 with value: 0.6887188408519016. regularization_factors, val_score: 0.688719: 50%|##### | 10/20 [00:11<00:11, 1.14s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013835 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515 Early stopping, best iteration is: [49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719 regularization_factors, val_score: 0.688719: 55%|#####5 | 11/20 [00:12<00:09, 1.07s/it][I 2020-09-27 04:50:02,068] Trial 53 finished with value: 0.688718840880323 and parameters: {'lambda_l1': 1.2436063977852807e-06, 'lambda_l2': 5.416531879194707e-07}. Best is trial 48 with value: 0.6887188408519016. regularization_factors, val_score: 0.688719: 55%|#####5 | 11/20 [00:12<00:09, 1.07s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.020591 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515 Early stopping, best iteration is: [49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719 regularization_factors, val_score: 0.688719: 60%|###### | 12/20 [00:13<00:08, 1.03s/it][I 2020-09-27 04:50:03,019] Trial 54 finished with value: 0.6887188408518048 and parameters: {'lambda_l1': 1.001927452586009e-08, 'lambda_l2': 1.0173335682514728e-08}. Best is trial 54 with value: 0.6887188408518048. regularization_factors, val_score: 0.688719: 60%|###### | 12/20 [00:13<00:08, 1.03s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013735 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515 Early stopping, best iteration is: [49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719 regularization_factors, val_score: 0.688719: 65%|######5 | 13/20 [00:14<00:07, 1.13s/it][I 2020-09-27 04:50:04,380] Trial 55 finished with value: 0.6887188408925199 and parameters: {'lambda_l1': 1.0246196926378488e-06, 'lambda_l2': 2.2586132878897466e-06}. Best is trial 54 with value: 0.6887188408518048. regularization_factors, val_score: 0.688719: 65%|######5 | 13/20 [00:14<00:07, 1.13s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.017187 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664481 valid's binary_logloss: 0.689645 Early stopping, best iteration is: [49] train's binary_logloss: 0.675961 valid's binary_logloss: 0.689284 regularization_factors, val_score: 0.688719: 70%|####### | 14/20 [00:15<00:06, 1.08s/it][I 2020-09-27 04:50:05,336] Trial 56 finished with value: 0.6892843278626516 and parameters: {'lambda_l1': 0.04091667144851904, 'lambda_l2': 1.0750374226083052e-08}. Best is trial 54 with value: 0.6887188408518048. regularization_factors, val_score: 0.688719: 70%|####### | 14/20 [00:15<00:06, 1.08s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013760 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515 Early stopping, best iteration is: [49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719 regularization_factors, val_score: 0.688719: 75%|#######5 | 15/20 [00:16<00:05, 1.04s/it][I 2020-09-27 04:50:06,270] Trial 57 finished with value: 0.6887188408561253 and parameters: {'lambda_l1': 1.5233181813455253e-07, 'lambda_l2': 1.8743684852533287e-07}. Best is trial 54 with value: 0.6887188408518048. regularization_factors, val_score: 0.688719: 75%|#######5 | 15/20 [00:16<00:05, 1.04s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.015130 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515 Early stopping, best iteration is: [49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719 regularization_factors, val_score: 0.688719: 80%|######## | 16/20 [00:19<00:06, 1.55s/it][I 2020-09-27 04:50:09,004] Trial 58 finished with value: 0.6887188422885628 and parameters: {'lambda_l1': 1.744760689790351e-07, 'lambda_l2': 0.00013088268252619333}. Best is trial 54 with value: 0.6887188408518048. regularization_factors, val_score: 0.688719: 80%|######## | 16/20 [00:19<00:06, 1.55s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.018473 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664768 valid's binary_logloss: 0.689166 Early stopping, best iteration is: [85] train's binary_logloss: 0.667917 valid's binary_logloss: 0.688911 regularization_factors, val_score: 0.688719: 85%|########5 | 17/20 [00:20<00:04, 1.45s/it][I 2020-09-27 04:50:10,235] Trial 59 finished with value: 0.6889111184760159 and parameters: {'lambda_l1': 0.00016154773417025348, 'lambda_l2': 0.9174839426871738}. Best is trial 54 with value: 0.6887188408518048. regularization_factors, val_score: 0.688719: 85%|########5 | 17/20 [00:20<00:04, 1.45s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.015601 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515 Early stopping, best iteration is: [49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719 regularization_factors, val_score: 0.688719: 90%|######### | 18/20 [00:21<00:02, 1.32s/it][I 2020-09-27 04:50:11,235] Trial 60 finished with value: 0.6887188408571195 and parameters: {'lambda_l1': 1.0344224061447487e-07, 'lambda_l2': 3.115824250207294e-07}. Best is trial 54 with value: 0.6887188408518048. regularization_factors, val_score: 0.688719: 90%|######### | 18/20 [00:21<00:02, 1.32s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016453 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515 Early stopping, best iteration is: [49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719 regularization_factors, val_score: 0.688719: 95%|#########5| 19/20 [00:23<00:01, 1.34s/it][I 2020-09-27 04:50:12,647] Trial 61 finished with value: 0.6887188408520247 and parameters: {'lambda_l1': 1.3726047963300931e-08, 'lambda_l2': 1.1171463229442362e-08}. Best is trial 54 with value: 0.6887188408518048. regularization_factors, val_score: 0.688719: 95%|#########5| 19/20 [00:23<00:01, 1.34s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.003196 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664324 valid's binary_logloss: 0.689515 Early stopping, best iteration is: [49] train's binary_logloss: 0.675893 valid's binary_logloss: 0.688719 regularization_factors, val_score: 0.688719: 100%|##########| 20/20 [00:24<00:00, 1.26s/it][I 2020-09-27 04:50:13,695] Trial 62 finished with value: 0.688718840851812 and parameters: {'lambda_l1': 1.0200361074017904e-08, 'lambda_l2': 1.1298563130408035e-08}. Best is trial 54 with value: 0.6887188408518048. regularization_factors, val_score: 0.688719: 100%|##########| 20/20 [00:24<00:00, 1.20s/it] min_data_in_leaf, val_score: 0.688719: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013241 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664243 valid's binary_logloss: 0.689415 Early stopping, best iteration is: [56] train's binary_logloss: 0.674277 valid's binary_logloss: 0.689398 min_data_in_leaf, val_score: 0.688719: 20%|## | 1/5 [00:00<00:03, 1.06it/s][I 2020-09-27 04:50:14,651] Trial 63 finished with value: 0.6893981290180272 and parameters: {'min_child_samples': 25}. Best is trial 63 with value: 0.6893981290180272. min_data_in_leaf, val_score: 0.688719: 20%|## | 1/5 [00:00<00:03, 1.06it/s][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012878 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663871 valid's binary_logloss: 0.689525 [200] train's binary_logloss: 0.643739 valid's binary_logloss: 0.690786 Early stopping, best iteration is: [124] train's binary_logloss: 0.658769 valid's binary_logloss: 0.689265 min_data_in_leaf, val_score: 0.688719: 40%|#### | 2/5 [00:02<00:03, 1.17s/it][I 2020-09-27 04:50:16,344] Trial 64 finished with value: 0.6892650557557729 and parameters: {'min_child_samples': 5}. Best is trial 64 with value: 0.6892650557557729. min_data_in_leaf, val_score: 0.688719: 40%|#### | 2/5 [00:02<00:03, 1.17s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013201 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664176 valid's binary_logloss: 0.689954 Early stopping, best iteration is: [55] train's binary_logloss: 0.674447 valid's binary_logloss: 0.689227 min_data_in_leaf, val_score: 0.688719: 60%|###### | 3/5 [00:03<00:02, 1.14s/it][I 2020-09-27 04:50:17,422] Trial 65 finished with value: 0.6892266404784428 and parameters: {'min_child_samples': 10}. Best is trial 65 with value: 0.6892266404784428. min_data_in_leaf, val_score: 0.688719: 60%|###### | 3/5 [00:03<00:02, 1.14s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013886 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664683 valid's binary_logloss: 0.689526 Early stopping, best iteration is: [69] train's binary_logloss: 0.67151 valid's binary_logloss: 0.689297 min_data_in_leaf, val_score: 0.688719: 80%|######## | 4/5 [00:04<00:01, 1.13s/it][I 2020-09-27 04:50:18,524] Trial 66 finished with value: 0.689296784041159 and parameters: {'min_child_samples': 50}. Best is trial 65 with value: 0.6892266404784428. min_data_in_leaf, val_score: 0.688719: 80%|######## | 4/5 [00:04<00:01, 1.13s/it][LightGBM] [Info] Number of positive: 46608, number of negative: 46417 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013526 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501027 -> initscore=0.004106 [LightGBM] [Info] Start training from score 0.004106 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665727 valid's binary_logloss: 0.689486 Early stopping, best iteration is: [74] train's binary_logloss: 0.67097 valid's binary_logloss: 0.689201 min_data_in_leaf, val_score: 0.688719: 100%|##########| 5/5 [00:06<00:00, 1.19s/it][I 2020-09-27 04:50:19,869] Trial 67 finished with value: 0.689200711916551 and parameters: {'min_child_samples': 100}. Best is trial 67 with value: 0.689200711916551. min_data_in_leaf, val_score: 0.688719: 100%|##########| 5/5 [00:06<00:00, 1.24s/it] Fold : 1 [I 2020-09-27 04:50:19,990] A new study created in memory with name: no-name-42491456-2d3a-4d80-ab38-ce544f64ff2c feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001692 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665024 valid's binary_logloss: 0.689794 Early stopping, best iteration is: [70] train's binary_logloss: 0.671608 valid's binary_logloss: 0.689532 feature_fraction, val_score: 0.689532: 14%|#4 | 1/7 [00:01<00:07, 1.22s/it][I 2020-09-27 04:50:21,234] Trial 0 finished with value: 0.689531715864851 and parameters: {'feature_fraction': 0.5}. Best is trial 0 with value: 0.689531715864851. feature_fraction, val_score: 0.689532: 14%|#4 | 1/7 [00:01<00:07, 1.22s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012878 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663919 valid's binary_logloss: 0.690116 Early stopping, best iteration is: [40] train's binary_logloss: 0.678175 valid's binary_logloss: 0.689468 feature_fraction, val_score: 0.689468: 29%|##8 | 2/7 [00:01<00:05, 1.09s/it][I 2020-09-27 04:50:22,004] Trial 1 finished with value: 0.6894679163318443 and parameters: {'feature_fraction': 0.6}. Best is trial 1 with value: 0.6894679163318443. feature_fraction, val_score: 0.689468: 29%|##8 | 2/7 [00:01<00:05, 1.09s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011687 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.66358 valid's binary_logloss: 0.690887 Early stopping, best iteration is: [35] train's binary_logloss: 0.679357 valid's binary_logloss: 0.690079 feature_fraction, val_score: 0.689468: 43%|####2 | 3/7 [00:02<00:03, 1.01it/s][I 2020-09-27 04:50:22,771] Trial 2 finished with value: 0.6900787550751419 and parameters: {'feature_fraction': 0.7}. Best is trial 1 with value: 0.6894679163318443. feature_fraction, val_score: 0.689468: 43%|####2 | 3/7 [00:02<00:03, 1.01it/s][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.003338 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.662262 valid's binary_logloss: 0.690303 Early stopping, best iteration is: [93] train's binary_logloss: 0.664122 valid's binary_logloss: 0.689979 feature_fraction, val_score: 0.689468: 57%|#####7 | 4/7 [00:04<00:03, 1.18s/it][I 2020-09-27 04:50:24,375] Trial 3 finished with value: 0.6899793275720499 and parameters: {'feature_fraction': 1.0}. Best is trial 1 with value: 0.6894679163318443. feature_fraction, val_score: 0.689468: 57%|#####7 | 4/7 [00:04<00:03, 1.18s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.021087 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.662515 valid's binary_logloss: 0.689981 Early stopping, best iteration is: [51] train's binary_logloss: 0.674455 valid's binary_logloss: 0.689748 feature_fraction, val_score: 0.689468: 71%|#######1 | 5/7 [00:05<00:02, 1.10s/it][I 2020-09-27 04:50:25,314] Trial 4 finished with value: 0.6897477331954484 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 1 with value: 0.6894679163318443. feature_fraction, val_score: 0.689468: 71%|#######1 | 5/7 [00:05<00:02, 1.10s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001352 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.666503 valid's binary_logloss: 0.689198 [200] train's binary_logloss: 0.647913 valid's binary_logloss: 0.690111 Early stopping, best iteration is: [141] train's binary_logloss: 0.658411 valid's binary_logloss: 0.689014 feature_fraction, val_score: 0.689014: 86%|########5 | 6/7 [00:06<00:01, 1.12s/it][I 2020-09-27 04:50:26,470] Trial 5 finished with value: 0.6890136743093267 and parameters: {'feature_fraction': 0.4}. Best is trial 5 with value: 0.6890136743093267. feature_fraction, val_score: 0.689014: 86%|########5 | 6/7 [00:06<00:01, 1.12s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.018804 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.66328 valid's binary_logloss: 0.690498 Early stopping, best iteration is: [29] train's binary_logloss: 0.681042 valid's binary_logloss: 0.689914 feature_fraction, val_score: 0.689014: 100%|##########| 7/7 [00:07<00:00, 1.02s/it][I 2020-09-27 04:50:27,266] Trial 6 finished with value: 0.6899136884834198 and parameters: {'feature_fraction': 0.8}. Best is trial 5 with value: 0.6890136743093267. feature_fraction, val_score: 0.689014: 100%|##########| 7/7 [00:07<00:00, 1.04s/it] num_leaves, val_score: 0.689014: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001398 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.59502 valid's binary_logloss: 0.693529 Early stopping, best iteration is: [25] train's binary_logloss: 0.660259 valid's binary_logloss: 0.690335 num_leaves, val_score: 0.689014: 5%|5 | 1/20 [00:01<00:30, 1.59s/it][I 2020-09-27 04:50:28,875] Trial 7 finished with value: 0.6903350331651174 and parameters: {'num_leaves': 152}. Best is trial 7 with value: 0.6903350331651174. num_leaves, val_score: 0.689014: 5%|5 | 1/20 [00:01<00:30, 1.59s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001113 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.621859 valid's binary_logloss: 0.693386 Early stopping, best iteration is: [24] train's binary_logloss: 0.669901 valid's binary_logloss: 0.690519 num_leaves, val_score: 0.689014: 10%|# | 2/20 [00:02<00:25, 1.40s/it][I 2020-09-27 04:50:29,827] Trial 8 finished with value: 0.6905190321264499 and parameters: {'num_leaves': 103}. Best is trial 7 with value: 0.6903350331651174. num_leaves, val_score: 0.689014: 10%|# | 2/20 [00:02<00:25, 1.40s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001303 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.607653 valid's binary_logloss: 0.691023 Early stopping, best iteration is: [31] train's binary_logloss: 0.658737 valid's binary_logloss: 0.690547 num_leaves, val_score: 0.689014: 15%|#5 | 3/20 [00:03<00:22, 1.30s/it][I 2020-09-27 04:50:30,899] Trial 9 finished with value: 0.6905470890370582 and parameters: {'num_leaves': 128}. Best is trial 7 with value: 0.6903350331651174. num_leaves, val_score: 0.689014: 15%|#5 | 3/20 [00:03<00:22, 1.30s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001257 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68554 valid's binary_logloss: 0.689548 [200] train's binary_logloss: 0.681807 valid's binary_logloss: 0.689004 [300] train's binary_logloss: 0.678526 valid's binary_logloss: 0.688996 Early stopping, best iteration is: [221] train's binary_logloss: 0.681102 valid's binary_logloss: 0.688886 num_leaves, val_score: 0.688886: 20%|## | 4/20 [00:05<00:23, 1.44s/it][I 2020-09-27 04:50:32,672] Trial 10 finished with value: 0.6888861673616007 and parameters: {'num_leaves': 6}. Best is trial 10 with value: 0.6888861673616007. num_leaves, val_score: 0.688886: 20%|## | 4/20 [00:05<00:23, 1.44s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001441 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.674935 valid's binary_logloss: 0.689151 [200] train's binary_logloss: 0.663169 valid's binary_logloss: 0.689994 Early stopping, best iteration is: [122] train's binary_logloss: 0.672063 valid's binary_logloss: 0.689042 num_leaves, val_score: 0.688886: 25%|##5 | 5/20 [00:06<00:19, 1.33s/it][I 2020-09-27 04:50:33,728] Trial 11 finished with value: 0.6890424580437142 and parameters: {'num_leaves': 19}. Best is trial 10 with value: 0.6888861673616007. num_leaves, val_score: 0.688886: 25%|##5 | 5/20 [00:06<00:19, 1.33s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001371 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688532 valid's binary_logloss: 0.690109 [200] train's binary_logloss: 0.686793 valid's binary_logloss: 0.689219 [300] train's binary_logloss: 0.685512 valid's binary_logloss: 0.68876 [400] train's binary_logloss: 0.684466 valid's binary_logloss: 0.688694 [500] train's binary_logloss: 0.68348 valid's binary_logloss: 0.6886 Early stopping, best iteration is: [438] train's binary_logloss: 0.684084 valid's binary_logloss: 0.68855 num_leaves, val_score: 0.688550: 30%|### | 6/20 [00:08<00:23, 1.66s/it][I 2020-09-27 04:50:36,153] Trial 12 finished with value: 0.6885499863950102 and parameters: {'num_leaves': 3}. Best is trial 12 with value: 0.6885499863950102. num_leaves, val_score: 0.688550: 30%|### | 6/20 [00:08<00:23, 1.66s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001342 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68554 valid's binary_logloss: 0.689548 [200] train's binary_logloss: 0.681807 valid's binary_logloss: 0.689004 [300] train's binary_logloss: 0.678526 valid's binary_logloss: 0.688996 Early stopping, best iteration is: [221] train's binary_logloss: 0.681102 valid's binary_logloss: 0.688886 num_leaves, val_score: 0.688550: 35%|###5 | 7/20 [00:10<00:20, 1.54s/it][I 2020-09-27 04:50:37,431] Trial 13 finished with value: 0.6888861673616007 and parameters: {'num_leaves': 6}. Best is trial 12 with value: 0.6885499863950102. num_leaves, val_score: 0.688550: 35%|###5 | 7/20 [00:10<00:20, 1.54s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000758 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.657781 valid's binary_logloss: 0.68976 Early stopping, best iteration is: [90] train's binary_logloss: 0.660403 valid's binary_logloss: 0.6895 num_leaves, val_score: 0.688550: 40%|#### | 8/20 [00:11<00:16, 1.36s/it][I 2020-09-27 04:50:38,375] Trial 14 finished with value: 0.6895000449518462 and parameters: {'num_leaves': 43}. Best is trial 12 with value: 0.6885499863950102. num_leaves, val_score: 0.688550: 40%|#### | 8/20 [00:11<00:16, 1.36s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000790 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.64493 valid's binary_logloss: 0.691503 Early stopping, best iteration is: [44] train's binary_logloss: 0.667465 valid's binary_logloss: 0.690163 num_leaves, val_score: 0.688550: 45%|####5 | 9/20 [00:11<00:13, 1.20s/it][I 2020-09-27 04:50:39,181] Trial 15 finished with value: 0.6901631360080164 and parameters: {'num_leaves': 64}. Best is trial 12 with value: 0.6885499863950102. num_leaves, val_score: 0.688550: 45%|####5 | 9/20 [00:11<00:13, 1.20s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000739 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.548818 valid's binary_logloss: 0.697085 Early stopping, best iteration is: [20] train's binary_logloss: 0.653082 valid's binary_logloss: 0.690873 num_leaves, val_score: 0.688550: 50%|##### | 10/20 [00:13<00:12, 1.22s/it][I 2020-09-27 04:50:40,456] Trial 16 finished with value: 0.6908727065690785 and parameters: {'num_leaves': 247}. Best is trial 12 with value: 0.6885499863950102. num_leaves, val_score: 0.688550: 50%|##### | 10/20 [00:13<00:12, 1.22s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000719 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.690028 valid's binary_logloss: 0.690957 [200] train's binary_logloss: 0.688879 valid's binary_logloss: 0.690086 [300] train's binary_logloss: 0.688169 valid's binary_logloss: 0.689501 [400] train's binary_logloss: 0.687668 valid's binary_logloss: 0.689075 [500] train's binary_logloss: 0.687288 valid's binary_logloss: 0.688845 [600] train's binary_logloss: 0.686991 valid's binary_logloss: 0.688707 [700] train's binary_logloss: 0.68675 valid's binary_logloss: 0.688619 [800] train's binary_logloss: 0.686547 valid's binary_logloss: 0.688562 [900] train's binary_logloss: 0.686373 valid's binary_logloss: 0.688526 [1000] train's binary_logloss: 0.686221 valid's binary_logloss: 0.688477 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.686221 valid's binary_logloss: 0.688477 num_leaves, val_score: 0.688477: 55%|#####5 | 11/20 [00:16<00:15, 1.75s/it][I 2020-09-27 04:50:43,436] Trial 17 finished with value: 0.6884769144658212 and parameters: {'num_leaves': 2}. Best is trial 17 with value: 0.6884769144658212. num_leaves, val_score: 0.688477: 55%|#####5 | 11/20 [00:16<00:15, 1.75s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000850 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.573685 valid's binary_logloss: 0.696971 Early stopping, best iteration is: [20] train's binary_logloss: 0.660038 valid's binary_logloss: 0.691352 num_leaves, val_score: 0.688477: 60%|###### | 12/20 [00:17<00:12, 1.57s/it][I 2020-09-27 04:50:44,593] Trial 18 finished with value: 0.6913522035050282 and parameters: {'num_leaves': 195}. Best is trial 17 with value: 0.6884769144658212. num_leaves, val_score: 0.688477: 60%|###### | 12/20 [00:17<00:12, 1.57s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000815 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.64326 valid's binary_logloss: 0.690631 Early stopping, best iteration is: [71] train's binary_logloss: 0.654351 valid's binary_logloss: 0.690055 num_leaves, val_score: 0.688477: 65%|######5 | 13/20 [00:18<00:09, 1.39s/it][I 2020-09-27 04:50:45,549] Trial 19 finished with value: 0.6900553800103405 and parameters: {'num_leaves': 66}. Best is trial 17 with value: 0.6884769144658212. num_leaves, val_score: 0.688477: 65%|######5 | 13/20 [00:18<00:09, 1.39s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000857 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68554 valid's binary_logloss: 0.689548 [200] train's binary_logloss: 0.681807 valid's binary_logloss: 0.689004 [300] train's binary_logloss: 0.678526 valid's binary_logloss: 0.688996 Early stopping, best iteration is: [221] train's binary_logloss: 0.681102 valid's binary_logloss: 0.688886 num_leaves, val_score: 0.688477: 70%|####### | 14/20 [00:19<00:07, 1.30s/it][I 2020-09-27 04:50:46,632] Trial 20 finished with value: 0.6888861673616007 and parameters: {'num_leaves': 6}. Best is trial 17 with value: 0.6884769144658212. num_leaves, val_score: 0.688477: 70%|####### | 14/20 [00:19<00:07, 1.30s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000852 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688532 valid's binary_logloss: 0.690109 [200] train's binary_logloss: 0.686793 valid's binary_logloss: 0.689219 [300] train's binary_logloss: 0.685512 valid's binary_logloss: 0.68876 [400] train's binary_logloss: 0.684466 valid's binary_logloss: 0.688694 [500] train's binary_logloss: 0.68348 valid's binary_logloss: 0.6886 Early stopping, best iteration is: [438] train's binary_logloss: 0.684084 valid's binary_logloss: 0.68855 num_leaves, val_score: 0.688477: 75%|#######5 | 15/20 [00:21<00:07, 1.42s/it][I 2020-09-27 04:50:48,340] Trial 21 finished with value: 0.6885499863950102 and parameters: {'num_leaves': 3}. Best is trial 17 with value: 0.6884769144658212. num_leaves, val_score: 0.688477: 75%|#######5 | 15/20 [00:21<00:07, 1.42s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000875 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.660592 valid's binary_logloss: 0.690578 Early stopping, best iteration is: [41] train's binary_logloss: 0.676299 valid's binary_logloss: 0.690019 num_leaves, val_score: 0.688477: 80%|######## | 16/20 [00:21<00:04, 1.22s/it][I 2020-09-27 04:50:49,088] Trial 22 finished with value: 0.6900188884483167 and parameters: {'num_leaves': 39}. Best is trial 17 with value: 0.6884769144658212. num_leaves, val_score: 0.688477: 80%|######## | 16/20 [00:21<00:04, 1.22s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000823 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.658652 valid's binary_logloss: 0.689977 Early stopping, best iteration is: [49] train's binary_logloss: 0.672748 valid's binary_logloss: 0.689541 num_leaves, val_score: 0.688477: 85%|########5 | 17/20 [00:22<00:03, 1.09s/it][I 2020-09-27 04:50:49,865] Trial 23 finished with value: 0.6895412052804837 and parameters: {'num_leaves': 42}. Best is trial 17 with value: 0.6884769144658212. num_leaves, val_score: 0.688477: 85%|########5 | 17/20 [00:22<00:03, 1.09s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000785 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688532 valid's binary_logloss: 0.690109 [200] train's binary_logloss: 0.686793 valid's binary_logloss: 0.689219 [300] train's binary_logloss: 0.685512 valid's binary_logloss: 0.68876 [400] train's binary_logloss: 0.684466 valid's binary_logloss: 0.688694 [500] train's binary_logloss: 0.68348 valid's binary_logloss: 0.6886 Early stopping, best iteration is: [438] train's binary_logloss: 0.684084 valid's binary_logloss: 0.68855 num_leaves, val_score: 0.688477: 90%|######### | 18/20 [00:24<00:02, 1.28s/it][I 2020-09-27 04:50:51,590] Trial 24 finished with value: 0.6885499863950102 and parameters: {'num_leaves': 3}. Best is trial 17 with value: 0.6884769144658212. num_leaves, val_score: 0.688477: 90%|######### | 18/20 [00:24<00:02, 1.28s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000830 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.633629 valid's binary_logloss: 0.690048 Early stopping, best iteration is: [41] train's binary_logloss: 0.66326 valid's binary_logloss: 0.689286 num_leaves, val_score: 0.688477: 95%|#########5| 19/20 [00:25<00:01, 1.16s/it][I 2020-09-27 04:50:52,488] Trial 25 finished with value: 0.689285779471649 and parameters: {'num_leaves': 81}. Best is trial 17 with value: 0.6884769144658212. num_leaves, val_score: 0.688477: 95%|#########5| 19/20 [00:25<00:01, 1.16s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001270 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.669687 valid's binary_logloss: 0.68995 Early stopping, best iteration is: [70] train's binary_logloss: 0.674818 valid's binary_logloss: 0.689845 num_leaves, val_score: 0.688477: 100%|##########| 20/20 [00:25<00:00, 1.04s/it][I 2020-09-27 04:50:53,247] Trial 26 finished with value: 0.6898451259704299 and parameters: {'num_leaves': 26}. Best is trial 17 with value: 0.6884769144658212. num_leaves, val_score: 0.688477: 100%|##########| 20/20 [00:25<00:00, 1.30s/it] bagging, val_score: 0.688477: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000794 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689916 valid's binary_logloss: 0.690872 [200] train's binary_logloss: 0.688688 valid's binary_logloss: 0.689969 [300] train's binary_logloss: 0.687926 valid's binary_logloss: 0.689254 [400] train's binary_logloss: 0.687375 valid's binary_logloss: 0.688976 [500] train's binary_logloss: 0.686964 valid's binary_logloss: 0.688668 [600] train's binary_logloss: 0.686637 valid's binary_logloss: 0.688603 [700] train's binary_logloss: 0.686361 valid's binary_logloss: 0.688503 [800] train's binary_logloss: 0.686138 valid's binary_logloss: 0.688427 [900] train's binary_logloss: 0.68594 valid's binary_logloss: 0.68828 Early stopping, best iteration is: [893] train's binary_logloss: 0.685953 valid's binary_logloss: 0.688266 bagging, val_score: 0.688266: 10%|# | 1/10 [00:03<00:29, 3.27s/it][I 2020-09-27 04:50:56,524] Trial 27 finished with value: 0.6882656720836928 and parameters: {'bagging_fraction': 0.9056687704253129, 'bagging_freq': 1}. Best is trial 27 with value: 0.6882656720836928. bagging, val_score: 0.688266: 10%|# | 1/10 [00:03<00:29, 3.27s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000821 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689949 valid's binary_logloss: 0.690893 [200] train's binary_logloss: 0.688731 valid's binary_logloss: 0.689897 [300] train's binary_logloss: 0.687963 valid's binary_logloss: 0.689438 [400] train's binary_logloss: 0.687422 valid's binary_logloss: 0.688923 [500] train's binary_logloss: 0.687012 valid's binary_logloss: 0.688674 [600] train's binary_logloss: 0.686692 valid's binary_logloss: 0.688589 [700] train's binary_logloss: 0.686428 valid's binary_logloss: 0.688421 [800] train's binary_logloss: 0.686204 valid's binary_logloss: 0.688439 [900] train's binary_logloss: 0.686006 valid's binary_logloss: 0.688398 Early stopping, best iteration is: [868] train's binary_logloss: 0.686066 valid's binary_logloss: 0.68835 bagging, val_score: 0.688266: 20%|## | 2/10 [00:06<00:25, 3.24s/it][I 2020-09-27 04:50:59,688] Trial 28 finished with value: 0.68834951195225 and parameters: {'bagging_fraction': 0.9341933335925237, 'bagging_freq': 1}. Best is trial 27 with value: 0.6882656720836928. bagging, val_score: 0.688266: 20%|## | 2/10 [00:06<00:25, 3.24s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000760 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689958 valid's binary_logloss: 0.690901 [200] train's binary_logloss: 0.688753 valid's binary_logloss: 0.689921 [300] train's binary_logloss: 0.687991 valid's binary_logloss: 0.68943 [400] train's binary_logloss: 0.687457 valid's binary_logloss: 0.688954 [500] train's binary_logloss: 0.687054 valid's binary_logloss: 0.688639 [600] train's binary_logloss: 0.686734 valid's binary_logloss: 0.688575 [700] train's binary_logloss: 0.686471 valid's binary_logloss: 0.688404 [800] train's binary_logloss: 0.686247 valid's binary_logloss: 0.688399 [900] train's binary_logloss: 0.686054 valid's binary_logloss: 0.688395 [1000] train's binary_logloss: 0.685884 valid's binary_logloss: 0.688386 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.685884 valid's binary_logloss: 0.688386 bagging, val_score: 0.688266: 30%|### | 3/10 [00:09<00:22, 3.22s/it][I 2020-09-27 04:51:02,884] Trial 29 finished with value: 0.6883861307718222 and parameters: {'bagging_fraction': 0.9472695270258076, 'bagging_freq': 1}. Best is trial 27 with value: 0.6882656720836928. bagging, val_score: 0.688266: 30%|### | 3/10 [00:09<00:22, 3.22s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000818 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689959 valid's binary_logloss: 0.690908 [200] train's binary_logloss: 0.688754 valid's binary_logloss: 0.689942 [300] train's binary_logloss: 0.68799 valid's binary_logloss: 0.689383 [400] train's binary_logloss: 0.687454 valid's binary_logloss: 0.688954 [500] train's binary_logloss: 0.687052 valid's binary_logloss: 0.688747 [600] train's binary_logloss: 0.686731 valid's binary_logloss: 0.68862 [700] train's binary_logloss: 0.68647 valid's binary_logloss: 0.688482 [800] train's binary_logloss: 0.686248 valid's binary_logloss: 0.68848 Early stopping, best iteration is: [725] train's binary_logloss: 0.686411 valid's binary_logloss: 0.68843 bagging, val_score: 0.688266: 40%|#### | 4/10 [00:12<00:18, 3.05s/it][I 2020-09-27 04:51:05,526] Trial 30 finished with value: 0.6884302974689744 and parameters: {'bagging_fraction': 0.9478078952145158, 'bagging_freq': 1}. Best is trial 27 with value: 0.6882656720836928. bagging, val_score: 0.688266: 40%|#### | 4/10 [00:12<00:18, 3.05s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000815 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689962 valid's binary_logloss: 0.690903 [200] train's binary_logloss: 0.688757 valid's binary_logloss: 0.690012 [300] train's binary_logloss: 0.688001 valid's binary_logloss: 0.689453 [400] train's binary_logloss: 0.687468 valid's binary_logloss: 0.688994 [500] train's binary_logloss: 0.687058 valid's binary_logloss: 0.688733 [600] train's binary_logloss: 0.686738 valid's binary_logloss: 0.688599 [700] train's binary_logloss: 0.686473 valid's binary_logloss: 0.688496 [800] train's binary_logloss: 0.686255 valid's binary_logloss: 0.688458 [900] train's binary_logloss: 0.686061 valid's binary_logloss: 0.688445 Early stopping, best iteration is: [867] train's binary_logloss: 0.686119 valid's binary_logloss: 0.688391 bagging, val_score: 0.688266: 50%|##### | 5/10 [00:15<00:15, 3.08s/it][I 2020-09-27 04:51:08,673] Trial 31 finished with value: 0.6883910883412938 and parameters: {'bagging_fraction': 0.9500225140884224, 'bagging_freq': 1}. Best is trial 27 with value: 0.6882656720836928. bagging, val_score: 0.688266: 50%|##### | 5/10 [00:15<00:15, 3.08s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000836 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689952 valid's binary_logloss: 0.690877 [200] train's binary_logloss: 0.688739 valid's binary_logloss: 0.689932 [300] train's binary_logloss: 0.687981 valid's binary_logloss: 0.689456 [400] train's binary_logloss: 0.687443 valid's binary_logloss: 0.688946 [500] train's binary_logloss: 0.687035 valid's binary_logloss: 0.688722 [600] train's binary_logloss: 0.686712 valid's binary_logloss: 0.688567 [700] train's binary_logloss: 0.686447 valid's binary_logloss: 0.688364 [800] train's binary_logloss: 0.686224 valid's binary_logloss: 0.688372 Early stopping, best iteration is: [701] train's binary_logloss: 0.686445 valid's binary_logloss: 0.688354 bagging, val_score: 0.688266: 60%|###### | 6/10 [00:18<00:11, 2.94s/it][I 2020-09-27 04:51:11,277] Trial 32 finished with value: 0.6883544364816402 and parameters: {'bagging_fraction': 0.939276221019466, 'bagging_freq': 1}. Best is trial 27 with value: 0.6882656720836928. bagging, val_score: 0.688266: 60%|###### | 6/10 [00:18<00:11, 2.94s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000816 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689945 valid's binary_logloss: 0.690873 [200] train's binary_logloss: 0.688736 valid's binary_logloss: 0.68993 [300] train's binary_logloss: 0.687975 valid's binary_logloss: 0.689441 [400] train's binary_logloss: 0.687436 valid's binary_logloss: 0.688977 [500] train's binary_logloss: 0.687029 valid's binary_logloss: 0.68869 [600] train's binary_logloss: 0.68671 valid's binary_logloss: 0.688599 [700] train's binary_logloss: 0.686444 valid's binary_logloss: 0.688457 [800] train's binary_logloss: 0.686223 valid's binary_logloss: 0.688493 Early stopping, best iteration is: [700] train's binary_logloss: 0.686444 valid's binary_logloss: 0.688457 bagging, val_score: 0.688266: 70%|####### | 7/10 [00:20<00:08, 2.83s/it][I 2020-09-27 04:51:13,857] Trial 33 finished with value: 0.6884573747067545 and parameters: {'bagging_fraction': 0.9386192260753653, 'bagging_freq': 1}. Best is trial 27 with value: 0.6882656720836928. bagging, val_score: 0.688266: 70%|####### | 7/10 [00:20<00:08, 2.83s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000907 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689934 valid's binary_logloss: 0.690812 [200] train's binary_logloss: 0.688717 valid's binary_logloss: 0.689892 [300] train's binary_logloss: 0.687955 valid's binary_logloss: 0.689321 [400] train's binary_logloss: 0.687414 valid's binary_logloss: 0.688937 [500] train's binary_logloss: 0.686999 valid's binary_logloss: 0.688627 [600] train's binary_logloss: 0.686678 valid's binary_logloss: 0.68863 [700] train's binary_logloss: 0.686413 valid's binary_logloss: 0.688396 [800] train's binary_logloss: 0.68619 valid's binary_logloss: 0.688375 [900] train's binary_logloss: 0.685991 valid's binary_logloss: 0.688321 Early stopping, best iteration is: [867] train's binary_logloss: 0.686053 valid's binary_logloss: 0.688263 bagging, val_score: 0.688263: 80%|######## | 8/10 [00:23<00:05, 2.91s/it][I 2020-09-27 04:51:16,952] Trial 34 finished with value: 0.6882634163638113 and parameters: {'bagging_fraction': 0.9270871288128361, 'bagging_freq': 1}. Best is trial 34 with value: 0.6882634163638113. bagging, val_score: 0.688263: 80%|######## | 8/10 [00:23<00:05, 2.91s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000826 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627 [200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877 [300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924 [400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859 [500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858 [600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462 [700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201 Early stopping, best iteration is: [699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186 bagging, val_score: 0.688186: 90%|######### | 9/10 [00:26<00:02, 2.83s/it][I 2020-09-27 04:51:19,581] Trial 35 finished with value: 0.6881860644541565 and parameters: {'bagging_fraction': 0.797756062798586, 'bagging_freq': 1}. Best is trial 35 with value: 0.6881860644541565. bagging, val_score: 0.688186: 90%|######### | 9/10 [00:26<00:02, 2.83s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000754 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689842 valid's binary_logloss: 0.690437 [200] train's binary_logloss: 0.688572 valid's binary_logloss: 0.689737 [300] train's binary_logloss: 0.687777 valid's binary_logloss: 0.68909 [400] train's binary_logloss: 0.687207 valid's binary_logloss: 0.688463 [500] train's binary_logloss: 0.686786 valid's binary_logloss: 0.688389 Early stopping, best iteration is: [463] train's binary_logloss: 0.686927 valid's binary_logloss: 0.688299 bagging, val_score: 0.688186: 100%|##########| 10/10 [00:28<00:00, 2.52s/it][I 2020-09-27 04:51:21,392] Trial 36 finished with value: 0.6882991425357883 and parameters: {'bagging_fraction': 0.7512163390719674, 'bagging_freq': 4}. Best is trial 35 with value: 0.6881860644541565. bagging, val_score: 0.688186: 100%|##########| 10/10 [00:28<00:00, 2.81s/it] feature_fraction_stage2, val_score: 0.688186: 0%| | 0/3 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000836 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689843 valid's binary_logloss: 0.690616 [200] train's binary_logloss: 0.688551 valid's binary_logloss: 0.689846 [300] train's binary_logloss: 0.687748 valid's binary_logloss: 0.689146 [400] train's binary_logloss: 0.687179 valid's binary_logloss: 0.688875 [500] train's binary_logloss: 0.686761 valid's binary_logloss: 0.688605 [600] train's binary_logloss: 0.686432 valid's binary_logloss: 0.68842 [700] train's binary_logloss: 0.686158 valid's binary_logloss: 0.688213 Early stopping, best iteration is: [699] train's binary_logloss: 0.686161 valid's binary_logloss: 0.68819 feature_fraction_stage2, val_score: 0.688186: 33%|###3 | 1/3 [00:02<00:05, 2.77s/it][I 2020-09-27 04:51:24,172] Trial 37 finished with value: 0.6881903702314308 and parameters: {'feature_fraction': 0.41600000000000004}. Best is trial 37 with value: 0.6881903702314308. feature_fraction_stage2, val_score: 0.688186: 33%|###3 | 1/3 [00:02<00:05, 2.77s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000871 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689828 valid's binary_logloss: 0.690608 [200] train's binary_logloss: 0.688553 valid's binary_logloss: 0.689888 [300] train's binary_logloss: 0.687744 valid's binary_logloss: 0.689142 [400] train's binary_logloss: 0.687173 valid's binary_logloss: 0.688816 [500] train's binary_logloss: 0.686758 valid's binary_logloss: 0.688522 [600] train's binary_logloss: 0.686423 valid's binary_logloss: 0.688429 [700] train's binary_logloss: 0.686151 valid's binary_logloss: 0.688287 Early stopping, best iteration is: [697] train's binary_logloss: 0.686159 valid's binary_logloss: 0.688254 feature_fraction_stage2, val_score: 0.688186: 67%|######6 | 2/3 [00:05<00:02, 2.75s/it][I 2020-09-27 04:51:26,877] Trial 38 finished with value: 0.6882536571755299 and parameters: {'feature_fraction': 0.44800000000000006}. Best is trial 37 with value: 0.6881903702314308. feature_fraction_stage2, val_score: 0.688186: 67%|######6 | 2/3 [00:05<00:02, 2.75s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000812 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689828 valid's binary_logloss: 0.690608 [200] train's binary_logloss: 0.688553 valid's binary_logloss: 0.689888 [300] train's binary_logloss: 0.687744 valid's binary_logloss: 0.689142 [400] train's binary_logloss: 0.687173 valid's binary_logloss: 0.688816 [500] train's binary_logloss: 0.686758 valid's binary_logloss: 0.688522 [600] train's binary_logloss: 0.686423 valid's binary_logloss: 0.688429 [700] train's binary_logloss: 0.686151 valid's binary_logloss: 0.688287 Early stopping, best iteration is: [697] train's binary_logloss: 0.686159 valid's binary_logloss: 0.688254 feature_fraction_stage2, val_score: 0.688186: 100%|##########| 3/3 [00:08<00:00, 2.74s/it][I 2020-09-27 04:51:29,591] Trial 39 finished with value: 0.6882536571755299 and parameters: {'feature_fraction': 0.48000000000000004}. Best is trial 37 with value: 0.6881903702314308. feature_fraction_stage2, val_score: 0.688186: 100%|##########| 3/3 [00:08<00:00, 2.73s/it] regularization_factors, val_score: 0.688186: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008765 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627 [200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877 [300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924 [400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859 [500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858 [600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462 [700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201 Early stopping, best iteration is: [699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186 regularization_factors, val_score: 0.688186: 5%|5 | 1/20 [00:02<00:48, 2.55s/it][I 2020-09-27 04:51:32,160] Trial 40 finished with value: 0.6881860967213117 and parameters: {'lambda_l1': 4.739533108597076e-06, 'lambda_l2': 0.011058449085898892}. Best is trial 40 with value: 0.6881860967213117. regularization_factors, val_score: 0.688186: 5%|5 | 1/20 [00:02<00:48, 2.55s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000761 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627 [200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877 [300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924 [400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859 [500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858 [600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462 [700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201 Early stopping, best iteration is: [699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186 regularization_factors, val_score: 0.688186: 10%|# | 2/20 [00:05<00:46, 2.59s/it][I 2020-09-27 04:51:34,825] Trial 41 finished with value: 0.6881861017696784 and parameters: {'lambda_l1': 2.6160452041031395e-06, 'lambda_l2': 0.012798063629265445}. Best is trial 40 with value: 0.6881860967213117. regularization_factors, val_score: 0.688186: 10%|# | 2/20 [00:05<00:46, 2.59s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000802 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627 [200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877 [300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924 [400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859 [500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858 [600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462 [700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201 Early stopping, best iteration is: [699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186 regularization_factors, val_score: 0.688186: 15%|#5 | 3/20 [00:07<00:44, 2.59s/it][I 2020-09-27 04:51:37,430] Trial 42 finished with value: 0.6881860972873897 and parameters: {'lambda_l1': 2.1069017715929826e-06, 'lambda_l2': 0.011265308763779644}. Best is trial 40 with value: 0.6881860967213117. regularization_factors, val_score: 0.688186: 15%|#5 | 3/20 [00:07<00:44, 2.59s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000812 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627 [200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877 [300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924 [400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859 [500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858 [600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462 [700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201 Early stopping, best iteration is: [699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186 regularization_factors, val_score: 0.688186: 20%|## | 4/20 [00:10<00:42, 2.64s/it][I 2020-09-27 04:51:40,183] Trial 43 finished with value: 0.6881861188350943 and parameters: {'lambda_l1': 2.871077466536222e-06, 'lambda_l2': 0.01863739043244412}. Best is trial 40 with value: 0.6881860967213117. regularization_factors, val_score: 0.688186: 20%|## | 4/20 [00:10<00:42, 2.64s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000761 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627 [200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877 [300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924 [400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859 [500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858 [600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462 [700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201 Early stopping, best iteration is: [699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186 regularization_factors, val_score: 0.688186: 25%|##5 | 5/20 [00:13<00:39, 2.65s/it][I 2020-09-27 04:51:42,840] Trial 44 finished with value: 0.6881860987513834 and parameters: {'lambda_l1': 1.3228363853250858e-06, 'lambda_l2': 0.011769235325842804}. Best is trial 40 with value: 0.6881860967213117. regularization_factors, val_score: 0.688186: 25%|##5 | 5/20 [00:13<00:39, 2.65s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000756 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627 [200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877 [300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924 [400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859 [500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858 [600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462 [700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201 Early stopping, best iteration is: [699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186 regularization_factors, val_score: 0.688186: 30%|### | 6/20 [00:15<00:37, 2.66s/it][I 2020-09-27 04:51:45,544] Trial 45 finished with value: 0.6881861035517631 and parameters: {'lambda_l1': 1.083504577056653e-06, 'lambda_l2': 0.013413611446410094}. Best is trial 40 with value: 0.6881860967213117. regularization_factors, val_score: 0.688186: 30%|### | 6/20 [00:15<00:37, 2.66s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000815 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627 [200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877 [300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924 [400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859 [500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858 [600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462 [700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201 Early stopping, best iteration is: [699] train's binary_logloss: 0.686174 valid's binary_logloss: 0.688186 regularization_factors, val_score: 0.688186: 35%|###5 | 7/20 [00:18<00:34, 2.64s/it][I 2020-09-27 04:51:48,142] Trial 46 finished with value: 0.688186150142476 and parameters: {'lambda_l1': 3.3314910180930466e-06, 'lambda_l2': 0.029327749830602143}. Best is trial 40 with value: 0.6881860967213117. regularization_factors, val_score: 0.688186: 35%|###5 | 7/20 [00:18<00:34, 2.64s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000818 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627 [200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877 [300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924 [400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859 [500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858 [600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462 [700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201 Early stopping, best iteration is: [699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186 regularization_factors, val_score: 0.688186: 40%|#### | 8/20 [00:21<00:31, 2.65s/it][I 2020-09-27 04:51:50,799] Trial 47 finished with value: 0.6881860757482499 and parameters: {'lambda_l1': 2.3526621184608058e-06, 'lambda_l2': 0.003868598725242898}. Best is trial 47 with value: 0.6881860757482499. regularization_factors, val_score: 0.688186: 40%|#### | 8/20 [00:21<00:31, 2.65s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000808 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627 [200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877 [300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924 [400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859 [500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858 [600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462 [700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201 Early stopping, best iteration is: [699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186 regularization_factors, val_score: 0.688186: 45%|####5 | 9/20 [00:23<00:29, 2.65s/it][I 2020-09-27 04:51:53,467] Trial 48 finished with value: 0.6881860645612168 and parameters: {'lambda_l1': 3.130361329176902e-06, 'lambda_l2': 2.312087319678873e-05}. Best is trial 48 with value: 0.6881860645612168. regularization_factors, val_score: 0.688186: 45%|####5 | 9/20 [00:23<00:29, 2.65s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000816 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627 [200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877 [300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924 [400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859 [500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858 [600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462 [700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201 Early stopping, best iteration is: [699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186 regularization_factors, val_score: 0.688186: 50%|##### | 10/20 [00:26<00:26, 2.65s/it][I 2020-09-27 04:51:56,110] Trial 49 finished with value: 0.6881860867937425 and parameters: {'lambda_l1': 0.0016839304778010226, 'lambda_l2': 1.0716481604934569e-06}. Best is trial 48 with value: 0.6881860645612168. regularization_factors, val_score: 0.688186: 50%|##### | 10/20 [00:26<00:26, 2.65s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000807 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689854 valid's binary_logloss: 0.690628 [200] train's binary_logloss: 0.688573 valid's binary_logloss: 0.689878 [300] train's binary_logloss: 0.687778 valid's binary_logloss: 0.689241 [400] train's binary_logloss: 0.687211 valid's binary_logloss: 0.688892 [500] train's binary_logloss: 0.686781 valid's binary_logloss: 0.688591 [600] train's binary_logloss: 0.686449 valid's binary_logloss: 0.688474 [700] train's binary_logloss: 0.686174 valid's binary_logloss: 0.688261 [800] train's binary_logloss: 0.685939 valid's binary_logloss: 0.688313 Early stopping, best iteration is: [739] train's binary_logloss: 0.686081 valid's binary_logloss: 0.688233 regularization_factors, val_score: 0.688186: 55%|#####5 | 11/20 [00:29<00:24, 2.69s/it][I 2020-09-27 04:51:58,894] Trial 50 finished with value: 0.6882329753342665 and parameters: {'lambda_l1': 0.09692234973226965, 'lambda_l2': 8.775766596471004e-07}. Best is trial 48 with value: 0.6881860645612168. regularization_factors, val_score: 0.688186: 55%|#####5 | 11/20 [00:29<00:24, 2.69s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000815 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627 [200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877 [300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924 [400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859 [500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858 [600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462 [700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201 Early stopping, best iteration is: [699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186 regularization_factors, val_score: 0.688186: 60%|###### | 12/20 [00:31<00:21, 2.68s/it][I 2020-09-27 04:52:01,543] Trial 51 finished with value: 0.6881861075194643 and parameters: {'lambda_l1': 0.0032449073251714445, 'lambda_l2': 3.2716223509922217e-06}. Best is trial 48 with value: 0.6881860645612168. regularization_factors, val_score: 0.688186: 60%|###### | 12/20 [00:31<00:21, 2.68s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000769 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627 [200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877 [300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924 [400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859 [500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858 [600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462 [700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201 Early stopping, best iteration is: [699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186 regularization_factors, val_score: 0.688186: 65%|######5 | 13/20 [00:34<00:18, 2.67s/it][I 2020-09-27 04:52:04,186] Trial 52 finished with value: 0.6881860698451091 and parameters: {'lambda_l1': 0.0003848047657014028, 'lambda_l2': 9.591675343195672e-05}. Best is trial 48 with value: 0.6881860645612168. regularization_factors, val_score: 0.688186: 65%|######5 | 13/20 [00:34<00:18, 2.67s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000847 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627 [200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877 [300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924 [400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859 [500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858 [600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462 [700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201 Early stopping, best iteration is: [699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186 regularization_factors, val_score: 0.688186: 70%|####### | 14/20 [00:37<00:15, 2.65s/it][I 2020-09-27 04:52:06,789] Trial 53 finished with value: 0.6881860703621502 and parameters: {'lambda_l1': 0.0004165194013850262, 'lambda_l2': 0.0001302229603956366}. Best is trial 48 with value: 0.6881860645612168. regularization_factors, val_score: 0.688186: 70%|####### | 14/20 [00:37<00:15, 2.65s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000813 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627 [200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877 [300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924 [400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859 [500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858 [600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462 [700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201 Early stopping, best iteration is: [699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186 regularization_factors, val_score: 0.688186: 75%|#######5 | 15/20 [00:42<00:17, 3.49s/it][I 2020-09-27 04:52:12,308] Trial 54 finished with value: 0.6881860706975759 and parameters: {'lambda_l1': 0.00045861316138919045, 'lambda_l2': 5.257878461523017e-05}. Best is trial 48 with value: 0.6881860645612168. regularization_factors, val_score: 0.688186: 75%|#######5 | 15/20 [00:42<00:17, 3.49s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001772 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627 [200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877 [300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924 [400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859 [500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858 [600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462 [700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201 Early stopping, best iteration is: [699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186 regularization_factors, val_score: 0.688186: 80%|######## | 16/20 [00:47<00:15, 3.79s/it][I 2020-09-27 04:52:16,741] Trial 55 finished with value: 0.6881860672245419 and parameters: {'lambda_l1': 0.00019089434047484262, 'lambda_l2': 8.205565377180615e-05}. Best is trial 48 with value: 0.6881860645612168. regularization_factors, val_score: 0.688186: 80%|######## | 16/20 [00:47<00:15, 3.79s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000880 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627 [200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877 [300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924 [400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859 [500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858 [600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462 [700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201 Early stopping, best iteration is: [699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186 regularization_factors, val_score: 0.688186: 85%|########5 | 17/20 [00:50<00:10, 3.55s/it][I 2020-09-27 04:52:19,720] Trial 56 finished with value: 0.6881860669984095 and parameters: {'lambda_l1': 0.00017461333741205492, 'lambda_l2': 7.853596223254834e-05}. Best is trial 48 with value: 0.6881860645612168. regularization_factors, val_score: 0.688186: 85%|########5 | 17/20 [00:50<00:10, 3.55s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008942 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627 [200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877 [300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924 [400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859 [500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858 [600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462 [700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201 Early stopping, best iteration is: [699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186 regularization_factors, val_score: 0.688186: 90%|######### | 18/20 [00:53<00:06, 3.38s/it][I 2020-09-27 04:52:22,709] Trial 57 finished with value: 0.6881860668888051 and parameters: {'lambda_l1': 0.0001634804643071105, 'lambda_l2': 9.030958952909047e-05}. Best is trial 48 with value: 0.6881860645612168. regularization_factors, val_score: 0.688186: 90%|######### | 18/20 [00:53<00:06, 3.38s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000820 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627 [200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877 [300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924 [400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859 [500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858 [600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462 [700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201 Early stopping, best iteration is: [699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186 regularization_factors, val_score: 0.688186: 95%|#########5| 19/20 [00:56<00:03, 3.29s/it][I 2020-09-27 04:52:25,792] Trial 58 finished with value: 0.6881860653671973 and parameters: {'lambda_l1': 5.5880426121768656e-05, 'lambda_l2': 5.7285984514290246e-05}. Best is trial 48 with value: 0.6881860645612168. regularization_factors, val_score: 0.688186: 95%|#########5| 19/20 [00:56<00:03, 3.29s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001313 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627 [200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877 [300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924 [400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859 [500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.68858 [600] train's binary_logloss: 0.686448 valid's binary_logloss: 0.688462 [700] train's binary_logloss: 0.686171 valid's binary_logloss: 0.688201 Early stopping, best iteration is: [699] train's binary_logloss: 0.686173 valid's binary_logloss: 0.688186 regularization_factors, val_score: 0.688186: 100%|##########| 20/20 [00:59<00:00, 3.27s/it][I 2020-09-27 04:52:28,997] Trial 59 finished with value: 0.6881860649508688 and parameters: {'lambda_l1': 3.502289401416132e-05, 'lambda_l2': 1.1183876599807051e-05}. Best is trial 48 with value: 0.6881860645612168. regularization_factors, val_score: 0.688186: 100%|##########| 20/20 [00:59<00:00, 2.97s/it] min_data_in_leaf, val_score: 0.688186: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000910 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627 [200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877 [300] train's binary_logloss: 0.687776 valid's binary_logloss: 0.68924 [400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688859 [500] train's binary_logloss: 0.686783 valid's binary_logloss: 0.688582 [600] train's binary_logloss: 0.686449 valid's binary_logloss: 0.688469 [700] train's binary_logloss: 0.686176 valid's binary_logloss: 0.688262 Early stopping, best iteration is: [699] train's binary_logloss: 0.686178 valid's binary_logloss: 0.688248 min_data_in_leaf, val_score: 0.688186: 20%|## | 1/5 [00:02<00:11, 2.91s/it][I 2020-09-27 04:52:31,924] Trial 60 finished with value: 0.6882475377052306 and parameters: {'min_child_samples': 25}. Best is trial 60 with value: 0.6882475377052306. min_data_in_leaf, val_score: 0.688186: 20%|## | 1/5 [00:02<00:11, 2.91s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000804 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627 [200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689877 [300] train's binary_logloss: 0.687769 valid's binary_logloss: 0.689167 [400] train's binary_logloss: 0.687205 valid's binary_logloss: 0.688804 [500] train's binary_logloss: 0.68679 valid's binary_logloss: 0.688551 [600] train's binary_logloss: 0.686455 valid's binary_logloss: 0.688496 [700] train's binary_logloss: 0.686182 valid's binary_logloss: 0.68822 Early stopping, best iteration is: [699] train's binary_logloss: 0.686183 valid's binary_logloss: 0.688205 min_data_in_leaf, val_score: 0.688186: 40%|#### | 2/5 [00:05<00:08, 2.90s/it][I 2020-09-27 04:52:34,798] Trial 61 finished with value: 0.688205087532443 and parameters: {'min_child_samples': 50}. Best is trial 61 with value: 0.688205087532443. min_data_in_leaf, val_score: 0.688186: 40%|#### | 2/5 [00:05<00:08, 2.90s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008806 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627 [200] train's binary_logloss: 0.688564 valid's binary_logloss: 0.689845 [300] train's binary_logloss: 0.687763 valid's binary_logloss: 0.689203 [400] train's binary_logloss: 0.687188 valid's binary_logloss: 0.688823 [500] train's binary_logloss: 0.686762 valid's binary_logloss: 0.688533 [600] train's binary_logloss: 0.686429 valid's binary_logloss: 0.688447 [700] train's binary_logloss: 0.686152 valid's binary_logloss: 0.688247 Early stopping, best iteration is: [699] train's binary_logloss: 0.686153 valid's binary_logloss: 0.688233 min_data_in_leaf, val_score: 0.688186: 60%|###### | 3/5 [00:08<00:05, 2.89s/it][I 2020-09-27 04:52:37,667] Trial 62 finished with value: 0.688232682196116 and parameters: {'min_child_samples': 10}. Best is trial 61 with value: 0.688205087532443. min_data_in_leaf, val_score: 0.688186: 60%|###### | 3/5 [00:08<00:05, 2.89s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000818 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689852 valid's binary_logloss: 0.690627 [200] train's binary_logloss: 0.688564 valid's binary_logloss: 0.689845 [300] train's binary_logloss: 0.687763 valid's binary_logloss: 0.689203 [400] train's binary_logloss: 0.687188 valid's binary_logloss: 0.688823 [500] train's binary_logloss: 0.686762 valid's binary_logloss: 0.688533 [600] train's binary_logloss: 0.686429 valid's binary_logloss: 0.688447 [700] train's binary_logloss: 0.686152 valid's binary_logloss: 0.688247 [800] train's binary_logloss: 0.68591 valid's binary_logloss: 0.688233 [900] train's binary_logloss: 0.685702 valid's binary_logloss: 0.688215 [1000] train's binary_logloss: 0.685499 valid's binary_logloss: 0.688212 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.685499 valid's binary_logloss: 0.688212 min_data_in_leaf, val_score: 0.688186: 80%|######## | 4/5 [00:12<00:03, 3.15s/it][I 2020-09-27 04:52:41,438] Trial 63 finished with value: 0.6882121446078117 and parameters: {'min_child_samples': 5}. Best is trial 61 with value: 0.688205087532443. min_data_in_leaf, val_score: 0.688186: 80%|######## | 4/5 [00:12<00:03, 3.15s/it][LightGBM] [Info] Number of positive: 46746, number of negative: 46279 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000805 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93025, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502510 -> initscore=0.010040 [LightGBM] [Info] Start training from score 0.010040 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689902 valid's binary_logloss: 0.690519 [200] train's binary_logloss: 0.68864 valid's binary_logloss: 0.68968 [300] train's binary_logloss: 0.687867 valid's binary_logloss: 0.688977 [400] train's binary_logloss: 0.687339 valid's binary_logloss: 0.688725 [500] train's binary_logloss: 0.686927 valid's binary_logloss: 0.688484 [600] train's binary_logloss: 0.686611 valid's binary_logloss: 0.688434 Early stopping, best iteration is: [531] train's binary_logloss: 0.68682 valid's binary_logloss: 0.688362 min_data_in_leaf, val_score: 0.688186: 100%|##########| 5/5 [00:15<00:00, 3.03s/it][I 2020-09-27 04:52:44,170] Trial 64 finished with value: 0.6883621338709048 and parameters: {'min_child_samples': 100}. Best is trial 61 with value: 0.688205087532443. min_data_in_leaf, val_score: 0.688186: 100%|##########| 5/5 [00:15<00:00, 3.03s/it] Fold : 2 [I 2020-09-27 04:52:44,427] A new study created in memory with name: no-name-4beb476a-e1cd-441a-9d78-1d4a1c5ec6fe feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010680 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663809 valid's binary_logloss: 0.689691 Early stopping, best iteration is: [55] train's binary_logloss: 0.674139 valid's binary_logloss: 0.689343 feature_fraction, val_score: 0.689343: 14%|#4 | 1/7 [00:01<00:09, 1.55s/it][I 2020-09-27 04:52:45,997] Trial 0 finished with value: 0.689343019442042 and parameters: {'feature_fraction': 0.7}. Best is trial 0 with value: 0.689343019442042. feature_fraction, val_score: 0.689343: 14%|#4 | 1/7 [00:01<00:09, 1.55s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008342 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664318 valid's binary_logloss: 0.688991 Early stopping, best iteration is: [63] train's binary_logloss: 0.672462 valid's binary_logloss: 0.688554 feature_fraction, val_score: 0.688554: 29%|##8 | 2/7 [00:03<00:07, 1.58s/it][I 2020-09-27 04:52:47,648] Trial 1 finished with value: 0.6885538121639901 and parameters: {'feature_fraction': 0.6}. Best is trial 1 with value: 0.6885538121639901. feature_fraction, val_score: 0.688554: 29%|##8 | 2/7 [00:03<00:07, 1.58s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001457 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.666146 valid's binary_logloss: 0.68872 Early stopping, best iteration is: [99] train's binary_logloss: 0.66631 valid's binary_logloss: 0.688654 feature_fraction, val_score: 0.688554: 43%|####2 | 3/7 [00:05<00:06, 1.68s/it][I 2020-09-27 04:52:49,556] Trial 2 finished with value: 0.6886536923224554 and parameters: {'feature_fraction': 0.4}. Best is trial 1 with value: 0.6885538121639901. feature_fraction, val_score: 0.688554: 43%|####2 | 3/7 [00:05<00:06, 1.68s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001582 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811 [200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036 Early stopping, best iteration is: [128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306 feature_fraction, val_score: 0.688306: 57%|#####7 | 4/7 [00:06<00:04, 1.56s/it][I 2020-09-27 04:52:50,831] Trial 3 finished with value: 0.6883061543757897 and parameters: {'feature_fraction': 0.5}. Best is trial 3 with value: 0.6883061543757897. feature_fraction, val_score: 0.688306: 57%|#####7 | 4/7 [00:06<00:04, 1.56s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013980 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663334 valid's binary_logloss: 0.689928 Early stopping, best iteration is: [39] train's binary_logloss: 0.678023 valid's binary_logloss: 0.689633 feature_fraction, val_score: 0.688306: 71%|#######1 | 5/7 [00:07<00:02, 1.33s/it][I 2020-09-27 04:52:51,651] Trial 4 finished with value: 0.6896333450899451 and parameters: {'feature_fraction': 0.8}. Best is trial 3 with value: 0.6883061543757897. feature_fraction, val_score: 0.688306: 71%|#######1 | 5/7 [00:07<00:02, 1.33s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.003355 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.662709 valid's binary_logloss: 0.690225 Early stopping, best iteration is: [72] train's binary_logloss: 0.669305 valid's binary_logloss: 0.689574 feature_fraction, val_score: 0.688306: 86%|########5 | 6/7 [00:08<00:01, 1.26s/it][I 2020-09-27 04:52:52,738] Trial 5 finished with value: 0.6895740757388882 and parameters: {'feature_fraction': 1.0}. Best is trial 3 with value: 0.6883061543757897. feature_fraction, val_score: 0.688306: 86%|########5 | 6/7 [00:08<00:01, 1.26s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001567 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663043 valid's binary_logloss: 0.689123 Early stopping, best iteration is: [74] train's binary_logloss: 0.668993 valid's binary_logloss: 0.688769 feature_fraction, val_score: 0.688306: 100%|##########| 7/7 [00:09<00:00, 1.20s/it][I 2020-09-27 04:52:53,795] Trial 6 finished with value: 0.6887687471521096 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 3 with value: 0.6883061543757897. feature_fraction, val_score: 0.688306: 100%|##########| 7/7 [00:09<00:00, 1.34s/it] num_leaves, val_score: 0.688306: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010710 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.642699 valid's binary_logloss: 0.689465 Early stopping, best iteration is: [60] train's binary_logloss: 0.658614 valid's binary_logloss: 0.688722 num_leaves, val_score: 0.688306: 5%|5 | 1/20 [00:01<00:19, 1.05s/it][I 2020-09-27 04:52:54,861] Trial 7 finished with value: 0.6887216519277777 and parameters: {'num_leaves': 63}. Best is trial 7 with value: 0.6887216519277777. num_leaves, val_score: 0.688306: 5%|5 | 1/20 [00:01<00:19, 1.05s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001033 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.541922 valid's binary_logloss: 0.696672 Early stopping, best iteration is: [17] train's binary_logloss: 0.656043 valid's binary_logloss: 0.691091 num_leaves, val_score: 0.688306: 10%|# | 2/20 [00:02<00:21, 1.17s/it][I 2020-09-27 04:52:56,310] Trial 8 finished with value: 0.6910909448392947 and parameters: {'num_leaves': 247}. Best is trial 7 with value: 0.6887216519277777. num_leaves, val_score: 0.688306: 10%|# | 2/20 [00:02<00:21, 1.17s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000948 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.661372 valid's binary_logloss: 0.689877 Early stopping, best iteration is: [75] train's binary_logloss: 0.667459 valid's binary_logloss: 0.689409 num_leaves, val_score: 0.688306: 15%|#5 | 3/20 [00:03<00:18, 1.09s/it][I 2020-09-27 04:52:57,203] Trial 9 finished with value: 0.6894087377385544 and parameters: {'num_leaves': 36}. Best is trial 7 with value: 0.6887216519277777. num_leaves, val_score: 0.688306: 15%|#5 | 3/20 [00:03<00:18, 1.09s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000877 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.543852 valid's binary_logloss: 0.693135 Early stopping, best iteration is: [20] train's binary_logloss: 0.651184 valid's binary_logloss: 0.690432 num_leaves, val_score: 0.688306: 20%|## | 4/20 [00:04<00:18, 1.17s/it][I 2020-09-27 04:52:58,582] Trial 10 finished with value: 0.6904319926657629 and parameters: {'num_leaves': 243}. Best is trial 7 with value: 0.6887216519277777. num_leaves, val_score: 0.688306: 20%|## | 4/20 [00:04<00:18, 1.17s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000959 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.589007 valid's binary_logloss: 0.693108 Early stopping, best iteration is: [37] train's binary_logloss: 0.644233 valid's binary_logloss: 0.690224 num_leaves, val_score: 0.688306: 25%|##5 | 5/20 [00:05<00:17, 1.19s/it][I 2020-09-27 04:52:59,794] Trial 11 finished with value: 0.6902240196479881 and parameters: {'num_leaves': 155}. Best is trial 7 with value: 0.6887216519277777. num_leaves, val_score: 0.688306: 25%|##5 | 5/20 [00:05<00:17, 1.19s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000952 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.596218 valid's binary_logloss: 0.693075 Early stopping, best iteration is: [34] train's binary_logloss: 0.650962 valid's binary_logloss: 0.689486 num_leaves, val_score: 0.688306: 30%|### | 6/20 [00:07<00:16, 1.17s/it][I 2020-09-27 04:53:00,935] Trial 12 finished with value: 0.689485631745598 and parameters: {'num_leaves': 141}. Best is trial 7 with value: 0.6887216519277777. num_leaves, val_score: 0.688306: 30%|### | 6/20 [00:07<00:16, 1.17s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000944 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.69009 valid's binary_logloss: 0.690453 [200] train's binary_logloss: 0.688904 valid's binary_logloss: 0.689731 [300] train's binary_logloss: 0.688162 valid's binary_logloss: 0.689318 [400] train's binary_logloss: 0.687642 valid's binary_logloss: 0.688959 [500] train's binary_logloss: 0.687262 valid's binary_logloss: 0.688713 [600] train's binary_logloss: 0.686964 valid's binary_logloss: 0.688609 [700] train's binary_logloss: 0.686721 valid's binary_logloss: 0.688505 [800] train's binary_logloss: 0.686514 valid's binary_logloss: 0.688439 [900] train's binary_logloss: 0.686339 valid's binary_logloss: 0.688402 [1000] train's binary_logloss: 0.686188 valid's binary_logloss: 0.688357 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.686188 valid's binary_logloss: 0.688357 num_leaves, val_score: 0.688306: 35%|###5 | 7/20 [00:10<00:23, 1.78s/it][I 2020-09-27 04:53:04,147] Trial 13 finished with value: 0.6883565901301231 and parameters: {'num_leaves': 2}. Best is trial 13 with value: 0.6883565901301231. num_leaves, val_score: 0.688306: 35%|###5 | 7/20 [00:10<00:23, 1.78s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001010 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.67672 valid's binary_logloss: 0.689306 [200] train's binary_logloss: 0.666267 valid's binary_logloss: 0.689375 Early stopping, best iteration is: [107] train's binary_logloss: 0.675908 valid's binary_logloss: 0.689164 num_leaves, val_score: 0.688306: 40%|#### | 8/20 [00:11<00:18, 1.53s/it][I 2020-09-27 04:53:05,090] Trial 14 finished with value: 0.6891638737730074 and parameters: {'num_leaves': 16}. Best is trial 13 with value: 0.6883565901301231. num_leaves, val_score: 0.688306: 40%|#### | 8/20 [00:11<00:18, 1.53s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000937 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.625772 valid's binary_logloss: 0.691507 Early stopping, best iteration is: [69] train's binary_logloss: 0.642008 valid's binary_logloss: 0.6905 num_leaves, val_score: 0.688306: 45%|####5 | 9/20 [00:12<00:15, 1.42s/it][I 2020-09-27 04:53:06,262] Trial 15 finished with value: 0.6904995062634406 and parameters: {'num_leaves': 90}. Best is trial 13 with value: 0.6883565901301231. num_leaves, val_score: 0.688306: 45%|####5 | 9/20 [00:12<00:15, 1.42s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001295 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.577877 valid's binary_logloss: 0.694164 Early stopping, best iteration is: [28] train's binary_logloss: 0.65064 valid's binary_logloss: 0.689958 num_leaves, val_score: 0.688306: 50%|##### | 10/20 [00:13<00:13, 1.38s/it][I 2020-09-27 04:53:07,528] Trial 16 finished with value: 0.6899575658354835 and parameters: {'num_leaves': 175}. Best is trial 13 with value: 0.6883565901301231. num_leaves, val_score: 0.688306: 50%|##### | 10/20 [00:13<00:13, 1.38s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000927 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.620139 valid's binary_logloss: 0.691195 Early stopping, best iteration is: [42] train's binary_logloss: 0.655396 valid's binary_logloss: 0.689705 num_leaves, val_score: 0.688306: 55%|#####5 | 11/20 [00:14<00:11, 1.26s/it][I 2020-09-27 04:53:08,520] Trial 17 finished with value: 0.6897051455072504 and parameters: {'num_leaves': 99}. Best is trial 13 with value: 0.6883565901301231. num_leaves, val_score: 0.688306: 55%|#####5 | 11/20 [00:14<00:11, 1.26s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007712 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.686259 valid's binary_logloss: 0.689494 [200] train's binary_logloss: 0.683128 valid's binary_logloss: 0.689035 [300] train's binary_logloss: 0.680538 valid's binary_logloss: 0.68913 Early stopping, best iteration is: [237] train's binary_logloss: 0.682127 valid's binary_logloss: 0.688927 num_leaves, val_score: 0.688306: 60%|###### | 12/20 [00:15<00:09, 1.22s/it][I 2020-09-27 04:53:09,653] Trial 18 finished with value: 0.6889270132361176 and parameters: {'num_leaves': 5}. Best is trial 13 with value: 0.6883565901301231. num_leaves, val_score: 0.688306: 60%|###### | 12/20 [00:15<00:09, 1.22s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001555 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.566389 valid's binary_logloss: 0.69472 Early stopping, best iteration is: [21] train's binary_logloss: 0.655891 valid's binary_logloss: 0.689891 num_leaves, val_score: 0.688306: 65%|######5 | 13/20 [00:17<00:08, 1.23s/it][I 2020-09-27 04:53:10,888] Trial 19 finished with value: 0.6898906845392911 and parameters: {'num_leaves': 199}. Best is trial 13 with value: 0.6883565901301231. num_leaves, val_score: 0.688306: 65%|######5 | 13/20 [00:17<00:08, 1.23s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000971 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.649368 valid's binary_logloss: 0.68967 Early stopping, best iteration is: [86] train's binary_logloss: 0.65419 valid's binary_logloss: 0.689079 num_leaves, val_score: 0.688306: 70%|####### | 14/20 [00:18<00:07, 1.17s/it][I 2020-09-27 04:53:11,922] Trial 20 finished with value: 0.689079067247648 and parameters: {'num_leaves': 53}. Best is trial 13 with value: 0.6883565901301231. num_leaves, val_score: 0.688306: 70%|####### | 14/20 [00:18<00:07, 1.17s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001094 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.616545 valid's binary_logloss: 0.691175 Early stopping, best iteration is: [24] train's binary_logloss: 0.668053 valid's binary_logloss: 0.690395 num_leaves, val_score: 0.688306: 75%|#######5 | 15/20 [00:19<00:05, 1.10s/it][I 2020-09-27 04:53:12,870] Trial 21 finished with value: 0.6903952269669797 and parameters: {'num_leaves': 105}. Best is trial 13 with value: 0.6883565901301231. num_leaves, val_score: 0.688306: 75%|#######5 | 15/20 [00:19<00:05, 1.10s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000957 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.560451 valid's binary_logloss: 0.695619 Early stopping, best iteration is: [21] train's binary_logloss: 0.654473 valid's binary_logloss: 0.690005 num_leaves, val_score: 0.688306: 80%|######## | 16/20 [00:20<00:04, 1.16s/it][I 2020-09-27 04:53:14,161] Trial 22 finished with value: 0.6900048045676193 and parameters: {'num_leaves': 209}. Best is trial 13 with value: 0.6883565901301231. num_leaves, val_score: 0.688306: 80%|######## | 16/20 [00:20<00:04, 1.16s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000964 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.683466 valid's binary_logloss: 0.688908 [200] train's binary_logloss: 0.678323 valid's binary_logloss: 0.688588 Early stopping, best iteration is: [128] train's binary_logloss: 0.681889 valid's binary_logloss: 0.688484 num_leaves, val_score: 0.688306: 85%|########5 | 17/20 [00:21<00:03, 1.08s/it][I 2020-09-27 04:53:15,053] Trial 23 finished with value: 0.68848442653872 and parameters: {'num_leaves': 8}. Best is trial 13 with value: 0.6883565901301231. num_leaves, val_score: 0.688306: 85%|########5 | 17/20 [00:21<00:03, 1.08s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000948 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.677505 valid's binary_logloss: 0.68888 [200] train's binary_logloss: 0.667867 valid's binary_logloss: 0.689195 Early stopping, best iteration is: [101] train's binary_logloss: 0.677397 valid's binary_logloss: 0.688863 num_leaves, val_score: 0.688306: 90%|######### | 18/20 [00:22<00:02, 1.01s/it][I 2020-09-27 04:53:15,914] Trial 24 finished with value: 0.6888633696872527 and parameters: {'num_leaves': 15}. Best is trial 13 with value: 0.6883565901301231. num_leaves, val_score: 0.688306: 90%|######### | 18/20 [00:22<00:02, 1.01s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000957 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687323 valid's binary_logloss: 0.689724 [200] train's binary_logloss: 0.684864 valid's binary_logloss: 0.68927 [300] train's binary_logloss: 0.682878 valid's binary_logloss: 0.688942 [400] train's binary_logloss: 0.681166 valid's binary_logloss: 0.688902 Early stopping, best iteration is: [348] train's binary_logloss: 0.682019 valid's binary_logloss: 0.688739 num_leaves, val_score: 0.688306: 95%|#########5| 19/20 [00:23<00:01, 1.18s/it][I 2020-09-27 04:53:17,497] Trial 25 finished with value: 0.688738728369025 and parameters: {'num_leaves': 4}. Best is trial 13 with value: 0.6883565901301231. num_leaves, val_score: 0.688306: 95%|#########5| 19/20 [00:23<00:01, 1.18s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000982 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.638699 valid's binary_logloss: 0.691078 Early stopping, best iteration is: [34] train's binary_logloss: 0.669389 valid's binary_logloss: 0.689921 num_leaves, val_score: 0.688306: 100%|##########| 20/20 [00:24<00:00, 1.09s/it][I 2020-09-27 04:53:18,377] Trial 26 finished with value: 0.6899211705265006 and parameters: {'num_leaves': 69}. Best is trial 13 with value: 0.6883565901301231. num_leaves, val_score: 0.688306: 100%|##########| 20/20 [00:24<00:00, 1.23s/it] bagging, val_score: 0.688306: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001241 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.66507 valid's binary_logloss: 0.68924 Early stopping, best iteration is: [81] train's binary_logloss: 0.669105 valid's binary_logloss: 0.689037 bagging, val_score: 0.688306: 10%|# | 1/10 [00:01<00:09, 1.05s/it][I 2020-09-27 04:53:19,441] Trial 27 finished with value: 0.6890373745262042 and parameters: {'bagging_fraction': 0.918366805969969, 'bagging_freq': 7}. Best is trial 27 with value: 0.6890373745262042. bagging, val_score: 0.688306: 10%|# | 1/10 [00:01<00:09, 1.05s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009692 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.666459 valid's binary_logloss: 0.689984 Early stopping, best iteration is: [75] train's binary_logloss: 0.671612 valid's binary_logloss: 0.68934 bagging, val_score: 0.688306: 20%|## | 2/10 [00:02<00:08, 1.03s/it][I 2020-09-27 04:53:20,408] Trial 28 finished with value: 0.6893396192520514 and parameters: {'bagging_fraction': 0.4219518026182199, 'bagging_freq': 1}. Best is trial 27 with value: 0.6890373745262042. bagging, val_score: 0.688306: 20%|## | 2/10 [00:02<00:08, 1.03s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000951 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.666862 valid's binary_logloss: 0.691464 Early stopping, best iteration is: [53] train's binary_logloss: 0.676595 valid's binary_logloss: 0.69025 bagging, val_score: 0.688306: 30%|### | 3/10 [00:02<00:06, 1.04it/s][I 2020-09-27 04:53:21,234] Trial 29 finished with value: 0.6902501595666304 and parameters: {'bagging_fraction': 0.4030288821219972, 'bagging_freq': 4}. Best is trial 27 with value: 0.6890373745262042. bagging, val_score: 0.688306: 30%|### | 3/10 [00:02<00:06, 1.04it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000959 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665081 valid's binary_logloss: 0.689749 [200] train's binary_logloss: 0.645221 valid's binary_logloss: 0.689436 Early stopping, best iteration is: [181] train's binary_logloss: 0.648839 valid's binary_logloss: 0.689363 bagging, val_score: 0.688306: 40%|#### | 4/10 [00:04<00:06, 1.10s/it][I 2020-09-27 04:53:22,634] Trial 30 finished with value: 0.6893631828354185 and parameters: {'bagging_fraction': 0.7296879278779264, 'bagging_freq': 1}. Best is trial 27 with value: 0.6890373745262042. bagging, val_score: 0.688306: 40%|#### | 4/10 [00:04<00:06, 1.10s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000952 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664945 valid's binary_logloss: 0.689235 Early stopping, best iteration is: [56] train's binary_logloss: 0.674807 valid's binary_logloss: 0.688819 bagging, val_score: 0.688306: 50%|##### | 5/10 [00:05<00:05, 1.04s/it][I 2020-09-27 04:53:23,533] Trial 31 finished with value: 0.6888194872886991 and parameters: {'bagging_fraction': 0.994076584593322, 'bagging_freq': 7}. Best is trial 31 with value: 0.6888194872886991. bagging, val_score: 0.688306: 50%|##### | 5/10 [00:05<00:05, 1.04s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000969 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665112 valid's binary_logloss: 0.689998 Early stopping, best iteration is: [64] train's binary_logloss: 0.673047 valid's binary_logloss: 0.689362 bagging, val_score: 0.688306: 60%|###### | 6/10 [00:06<00:03, 1.00it/s][I 2020-09-27 04:53:24,435] Trial 32 finished with value: 0.6893623976325989 and parameters: {'bagging_fraction': 0.6317840838824331, 'bagging_freq': 4}. Best is trial 31 with value: 0.6888194872886991. bagging, val_score: 0.688306: 60%|###### | 6/10 [00:06<00:03, 1.00it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000971 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665587 valid's binary_logloss: 0.689559 [200] train's binary_logloss: 0.646055 valid's binary_logloss: 0.689772 Early stopping, best iteration is: [147] train's binary_logloss: 0.656183 valid's binary_logloss: 0.689318 bagging, val_score: 0.688306: 70%|####### | 7/10 [00:07<00:03, 1.09s/it][I 2020-09-27 04:53:25,735] Trial 33 finished with value: 0.6893178393179011 and parameters: {'bagging_fraction': 0.6247158545663389, 'bagging_freq': 6}. Best is trial 31 with value: 0.6888194872886991. bagging, val_score: 0.688306: 70%|####### | 7/10 [00:07<00:03, 1.09s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000971 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665022 valid's binary_logloss: 0.68972 Early stopping, best iteration is: [60] train's binary_logloss: 0.673867 valid's binary_logloss: 0.689193 bagging, val_score: 0.688306: 80%|######## | 8/10 [00:08<00:02, 1.05s/it][I 2020-09-27 04:53:26,715] Trial 34 finished with value: 0.6891934518367359 and parameters: {'bagging_fraction': 0.8401040019057991, 'bagging_freq': 2}. Best is trial 31 with value: 0.6888194872886991. bagging, val_score: 0.688306: 80%|######## | 8/10 [00:08<00:02, 1.05s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000949 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665836 valid's binary_logloss: 0.690057 Early stopping, best iteration is: [26] train's binary_logloss: 0.683137 valid's binary_logloss: 0.69002 bagging, val_score: 0.688306: 90%|######### | 9/10 [00:09<00:00, 1.05it/s][I 2020-09-27 04:53:27,429] Trial 35 finished with value: 0.690019877609649 and parameters: {'bagging_fraction': 0.5293875757229788, 'bagging_freq': 3}. Best is trial 31 with value: 0.6888194872886991. bagging, val_score: 0.688306: 90%|######### | 9/10 [00:09<00:00, 1.05it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000988 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664989 valid's binary_logloss: 0.690033 Early stopping, best iteration is: [62] train's binary_logloss: 0.673472 valid's binary_logloss: 0.689508 bagging, val_score: 0.688306: 100%|##########| 10/10 [00:09<00:00, 1.06it/s][I 2020-09-27 04:53:28,340] Trial 36 finished with value: 0.6895082528824183 and parameters: {'bagging_fraction': 0.7844185449233542, 'bagging_freq': 5}. Best is trial 31 with value: 0.6888194872886991. bagging, val_score: 0.688306: 100%|##########| 10/10 [00:09<00:00, 1.00it/s] feature_fraction_stage2, val_score: 0.688306: 0%| | 0/6 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000900 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811 [200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036 Early stopping, best iteration is: [128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306 feature_fraction_stage2, val_score: 0.688306: 17%|#6 | 1/6 [00:01<00:05, 1.08s/it][I 2020-09-27 04:53:29,437] Trial 37 finished with value: 0.6883061543757897 and parameters: {'feature_fraction': 0.516}. Best is trial 37 with value: 0.6883061543757897. feature_fraction_stage2, val_score: 0.688306: 17%|#6 | 1/6 [00:01<00:05, 1.08s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000788 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664973 valid's binary_logloss: 0.689707 [200] train's binary_logloss: 0.646027 valid's binary_logloss: 0.690321 Early stopping, best iteration is: [117] train's binary_logloss: 0.661743 valid's binary_logloss: 0.689468 feature_fraction_stage2, val_score: 0.688306: 33%|###3 | 2/6 [00:02<00:04, 1.07s/it][I 2020-09-27 04:53:30,468] Trial 38 finished with value: 0.689467520428726 and parameters: {'feature_fraction': 0.45199999999999996}. Best is trial 37 with value: 0.6883061543757897. feature_fraction_stage2, val_score: 0.688306: 33%|###3 | 2/6 [00:02<00:04, 1.07s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001140 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.66468 valid's binary_logloss: 0.68972 Early stopping, best iteration is: [64] train's binary_logloss: 0.672487 valid's binary_logloss: 0.689477 feature_fraction_stage2, val_score: 0.688306: 50%|##### | 3/6 [00:02<00:03, 1.00s/it][I 2020-09-27 04:53:31,330] Trial 39 finished with value: 0.6894774349755185 and parameters: {'feature_fraction': 0.58}. Best is trial 37 with value: 0.6883061543757897. feature_fraction_stage2, val_score: 0.688306: 50%|##### | 3/6 [00:02<00:03, 1.00s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000843 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.66569 valid's binary_logloss: 0.688902 Early stopping, best iteration is: [65] train's binary_logloss: 0.673083 valid's binary_logloss: 0.688613 feature_fraction_stage2, val_score: 0.688306: 67%|######6 | 4/6 [00:03<00:01, 1.06it/s][I 2020-09-27 04:53:32,142] Trial 40 finished with value: 0.6886133673965851 and parameters: {'feature_fraction': 0.42}. Best is trial 37 with value: 0.6883061543757897. feature_fraction_stage2, val_score: 0.688306: 67%|######6 | 4/6 [00:03<00:01, 1.06it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000999 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664812 valid's binary_logloss: 0.688514 Early stopping, best iteration is: [92] train's binary_logloss: 0.66641 valid's binary_logloss: 0.688368 feature_fraction_stage2, val_score: 0.688306: 83%|########3 | 5/6 [00:04<00:00, 1.05it/s][I 2020-09-27 04:53:33,108] Trial 41 finished with value: 0.6883683592481646 and parameters: {'feature_fraction': 0.5479999999999999}. Best is trial 37 with value: 0.6883061543757897. feature_fraction_stage2, val_score: 0.688306: 83%|########3 | 5/6 [00:04<00:00, 1.05it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000975 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811 [200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036 Early stopping, best iteration is: [128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306 feature_fraction_stage2, val_score: 0.688306: 100%|##########| 6/6 [00:05<00:00, 1.01it/s][I 2020-09-27 04:53:34,199] Trial 42 finished with value: 0.6883061543757897 and parameters: {'feature_fraction': 0.484}. Best is trial 37 with value: 0.6883061543757897. feature_fraction_stage2, val_score: 0.688306: 100%|##########| 6/6 [00:05<00:00, 1.03it/s] regularization_factors, val_score: 0.688306: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000960 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.66526 valid's binary_logloss: 0.689325 Early stopping, best iteration is: [74] train's binary_logloss: 0.670714 valid's binary_logloss: 0.688927 regularization_factors, val_score: 0.688306: 5%|5 | 1/20 [00:00<00:16, 1.13it/s][I 2020-09-27 04:53:35,096] Trial 43 finished with value: 0.6889268680203512 and parameters: {'lambda_l1': 3.489683036490915e-05, 'lambda_l2': 0.03596499685458891}. Best is trial 43 with value: 0.6889268680203512. regularization_factors, val_score: 0.688306: 5%|5 | 1/20 [00:00<00:16, 1.13it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000943 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.668189 valid's binary_logloss: 0.688592 Early stopping, best iteration is: [76] train's binary_logloss: 0.672479 valid's binary_logloss: 0.688382 regularization_factors, val_score: 0.688306: 10%|# | 2/20 [00:01<00:16, 1.11it/s][I 2020-09-27 04:53:36,045] Trial 44 finished with value: 0.6883820148603388 and parameters: {'lambda_l1': 4.2910298459529574, 'lambda_l2': 4.4841181697512024e-08}. Best is trial 44 with value: 0.6883820148603388. regularization_factors, val_score: 0.688306: 10%|# | 2/20 [00:01<00:16, 1.11it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001239 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811 [200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036 Early stopping, best iteration is: [128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306 regularization_factors, val_score: 0.688306: 15%|#5 | 3/20 [00:02<00:16, 1.04it/s][I 2020-09-27 04:53:37,146] Trial 45 finished with value: 0.6883061543637848 and parameters: {'lambda_l1': 1.6123080908302226e-08, 'lambda_l2': 1.1923474224643993e-07}. Best is trial 45 with value: 0.6883061543637848. regularization_factors, val_score: 0.688306: 15%|#5 | 3/20 [00:02<00:16, 1.04it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001032 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811 [200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036 Early stopping, best iteration is: [128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306 regularization_factors, val_score: 0.688306: 20%|## | 4/20 [00:04<00:16, 1.01s/it][I 2020-09-27 04:53:38,271] Trial 46 finished with value: 0.6883061543726559 and parameters: {'lambda_l1': 1.4997584226077242e-08, 'lambda_l2': 1.9309033302845346e-08}. Best is trial 45 with value: 0.6883061543637848. regularization_factors, val_score: 0.688306: 20%|## | 4/20 [00:04<00:16, 1.01s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000984 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811 [200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036 Early stopping, best iteration is: [128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306 regularization_factors, val_score: 0.688306: 25%|##5 | 5/20 [00:05<00:15, 1.05s/it][I 2020-09-27 04:53:39,412] Trial 47 finished with value: 0.6883061543725224 and parameters: {'lambda_l1': 2.1537994864162705e-08, 'lambda_l2': 1.7013032790342447e-08}. Best is trial 45 with value: 0.6883061543637848. regularization_factors, val_score: 0.688306: 25%|##5 | 5/20 [00:05<00:15, 1.05s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000950 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811 [200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036 Early stopping, best iteration is: [128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306 regularization_factors, val_score: 0.688306: 30%|### | 6/20 [00:06<00:14, 1.06s/it][I 2020-09-27 04:53:40,511] Trial 48 finished with value: 0.6883061543697307 and parameters: {'lambda_l1': 4.5182334372625207e-08, 'lambda_l2': 1.1900726497967697e-08}. Best is trial 45 with value: 0.6883061543637848. regularization_factors, val_score: 0.688306: 30%|### | 6/20 [00:06<00:14, 1.06s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000959 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811 [200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036 Early stopping, best iteration is: [128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306 regularization_factors, val_score: 0.688306: 35%|###5 | 7/20 [00:07<00:13, 1.07s/it][I 2020-09-27 04:53:41,604] Trial 49 finished with value: 0.6883061543737345 and parameters: {'lambda_l1': 1.214350265927941e-08, 'lambda_l2': 1.243915863528419e-08}. Best is trial 45 with value: 0.6883061543637848. regularization_factors, val_score: 0.688306: 35%|###5 | 7/20 [00:07<00:13, 1.07s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000942 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811 [200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036 Early stopping, best iteration is: [128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306 regularization_factors, val_score: 0.688306: 40%|#### | 8/20 [00:08<00:12, 1.08s/it][I 2020-09-27 04:53:42,706] Trial 50 finished with value: 0.6883061543740199 and parameters: {'lambda_l1': 1.0018700830284507e-08, 'lambda_l2': 1.1810016327451072e-08}. Best is trial 45 with value: 0.6883061543637848. regularization_factors, val_score: 0.688306: 40%|#### | 8/20 [00:08<00:12, 1.08s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000961 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811 [200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036 Early stopping, best iteration is: [128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306 regularization_factors, val_score: 0.688306: 45%|####5 | 9/20 [00:09<00:12, 1.10s/it][I 2020-09-27 04:53:43,836] Trial 51 finished with value: 0.6883061543739415 and parameters: {'lambda_l1': 1.034818635491755e-08, 'lambda_l2': 1.1390198269989985e-08}. Best is trial 45 with value: 0.6883061543637848. regularization_factors, val_score: 0.688306: 45%|####5 | 9/20 [00:09<00:12, 1.10s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000968 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811 [200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036 Early stopping, best iteration is: [128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306 regularization_factors, val_score: 0.688306: 50%|##### | 10/20 [00:10<00:11, 1.12s/it][I 2020-09-27 04:53:45,004] Trial 52 finished with value: 0.6883061543730142 and parameters: {'lambda_l1': 1.67980865791993e-08, 'lambda_l2': 1.4273245882473324e-08}. Best is trial 45 with value: 0.6883061543637848. regularization_factors, val_score: 0.688306: 50%|##### | 10/20 [00:10<00:11, 1.12s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000947 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811 [200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036 Early stopping, best iteration is: [128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306 regularization_factors, val_score: 0.688306: 55%|#####5 | 11/20 [00:11<00:10, 1.11s/it][I 2020-09-27 04:53:46,102] Trial 53 finished with value: 0.6883061543727123 and parameters: {'lambda_l1': 1.54156161065481e-08, 'lambda_l2': 1.961580255914531e-08}. Best is trial 45 with value: 0.6883061543637848. regularization_factors, val_score: 0.688306: 55%|#####5 | 11/20 [00:11<00:10, 1.11s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001597 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811 [200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036 Early stopping, best iteration is: [128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306 regularization_factors, val_score: 0.688306: 60%|###### | 12/20 [00:12<00:08, 1.10s/it][I 2020-09-27 04:53:47,171] Trial 54 finished with value: 0.688306154369646 and parameters: {'lambda_l1': 2.6426720558168387e-08, 'lambda_l2': 3.5118264244271155e-08}. Best is trial 45 with value: 0.6883061543637848. regularization_factors, val_score: 0.688306: 60%|###### | 12/20 [00:12<00:08, 1.10s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000949 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811 [200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036 Early stopping, best iteration is: [128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306 regularization_factors, val_score: 0.688306: 65%|######5 | 13/20 [00:14<00:07, 1.10s/it][I 2020-09-27 04:53:48,280] Trial 55 finished with value: 0.6883061543276774 and parameters: {'lambda_l1': 6.8998257219367e-08, 'lambda_l2': 5.65558107194084e-07}. Best is trial 55 with value: 0.6883061543276774. regularization_factors, val_score: 0.688306: 65%|######5 | 13/20 [00:14<00:07, 1.10s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001012 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811 [200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036 Early stopping, best iteration is: [128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306 regularization_factors, val_score: 0.688306: 70%|####### | 14/20 [00:15<00:06, 1.13s/it][I 2020-09-27 04:53:49,474] Trial 56 finished with value: 0.6883061540275088 and parameters: {'lambda_l1': 7.521323057192409e-07, 'lambda_l2': 3.7320052285004596e-06}. Best is trial 56 with value: 0.6883061540275088. regularization_factors, val_score: 0.688306: 70%|####### | 14/20 [00:15<00:06, 1.13s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000962 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811 [200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036 Early stopping, best iteration is: [128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306 regularization_factors, val_score: 0.688306: 75%|#######5 | 15/20 [00:16<00:05, 1.13s/it][I 2020-09-27 04:53:50,589] Trial 57 finished with value: 0.6883061539600913 and parameters: {'lambda_l1': 1.1451365202695885e-06, 'lambda_l2': 4.1525999441678045e-06}. Best is trial 57 with value: 0.6883061539600913. regularization_factors, val_score: 0.688306: 75%|#######5 | 15/20 [00:16<00:05, 1.13s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000947 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811 [200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036 Early stopping, best iteration is: [128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306 regularization_factors, val_score: 0.688306: 80%|######## | 16/20 [00:17<00:04, 1.12s/it][I 2020-09-27 04:53:51,694] Trial 58 finished with value: 0.6883061537046381 and parameters: {'lambda_l1': 1.87654882489576e-06, 'lambda_l2': 6.539895147651026e-06}. Best is trial 58 with value: 0.6883061537046381. regularization_factors, val_score: 0.688306: 80%|######## | 16/20 [00:17<00:04, 1.12s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000888 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811 [200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036 Early stopping, best iteration is: [128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306 regularization_factors, val_score: 0.688306: 85%|########5 | 17/20 [00:18<00:03, 1.11s/it][I 2020-09-27 04:53:52,787] Trial 59 finished with value: 0.6883061539934588 and parameters: {'lambda_l1': 1.7458283390546473e-06, 'lambda_l2': 2.9537260784896182e-06}. Best is trial 58 with value: 0.6883061537046381. regularization_factors, val_score: 0.688306: 85%|########5 | 17/20 [00:18<00:03, 1.11s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000908 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811 [200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036 Early stopping, best iteration is: [128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306 regularization_factors, val_score: 0.688306: 90%|######### | 18/20 [00:19<00:02, 1.11s/it][I 2020-09-27 04:53:53,910] Trial 60 finished with value: 0.6883061538065175 and parameters: {'lambda_l1': 2.402166405130976e-06, 'lambda_l2': 4.623828082213961e-06}. Best is trial 58 with value: 0.6883061537046381. regularization_factors, val_score: 0.688306: 90%|######### | 18/20 [00:19<00:02, 1.11s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010324 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811 [200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036 Early stopping, best iteration is: [128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306 regularization_factors, val_score: 0.688306: 95%|#########5| 19/20 [00:20<00:01, 1.12s/it][I 2020-09-27 04:53:55,054] Trial 61 finished with value: 0.6883061537873091 and parameters: {'lambda_l1': 2.6852977733488266e-06, 'lambda_l2': 4.514170857027213e-06}. Best is trial 58 with value: 0.6883061537046381. regularization_factors, val_score: 0.688306: 95%|#########5| 19/20 [00:20<00:01, 1.12s/it][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012564 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665128 valid's binary_logloss: 0.688811 [200] train's binary_logloss: 0.646099 valid's binary_logloss: 0.689036 Early stopping, best iteration is: [128] train's binary_logloss: 0.659459 valid's binary_logloss: 0.688306 regularization_factors, val_score: 0.688306: 100%|##########| 20/20 [00:21<00:00, 1.10s/it][I 2020-09-27 04:53:56,095] Trial 62 finished with value: 0.6883061538017431 and parameters: {'lambda_l1': 2.099566024643849e-06, 'lambda_l2': 5.0450298575946105e-06}. Best is trial 58 with value: 0.6883061537046381. regularization_factors, val_score: 0.688306: 100%|##########| 20/20 [00:21<00:00, 1.09s/it] min_data_in_leaf, val_score: 0.688306: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000964 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665071 valid's binary_logloss: 0.68918 Early stopping, best iteration is: [66] train's binary_logloss: 0.672155 valid's binary_logloss: 0.688846 min_data_in_leaf, val_score: 0.688306: 20%|## | 1/5 [00:00<00:03, 1.20it/s][I 2020-09-27 04:53:56,941] Trial 63 finished with value: 0.6888461130563782 and parameters: {'min_child_samples': 10}. Best is trial 63 with value: 0.6888461130563782. min_data_in_leaf, val_score: 0.688306: 20%|## | 1/5 [00:00<00:03, 1.20it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001044 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.666223 valid's binary_logloss: 0.689352 Early stopping, best iteration is: [63] train's binary_logloss: 0.673794 valid's binary_logloss: 0.688948 min_data_in_leaf, val_score: 0.688306: 40%|#### | 2/5 [00:01<00:02, 1.20it/s][I 2020-09-27 04:53:57,783] Trial 64 finished with value: 0.6889480364726907 and parameters: {'min_child_samples': 100}. Best is trial 63 with value: 0.6888461130563782. min_data_in_leaf, val_score: 0.688306: 40%|#### | 2/5 [00:01<00:02, 1.20it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000895 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.6653 valid's binary_logloss: 0.688213 Early stopping, best iteration is: [98] train's binary_logloss: 0.665652 valid's binary_logloss: 0.688167 min_data_in_leaf, val_score: 0.688167: 60%|###### | 3/5 [00:02<00:01, 1.14it/s][I 2020-09-27 04:53:58,761] Trial 65 finished with value: 0.6881665249243496 and parameters: {'min_child_samples': 25}. Best is trial 65 with value: 0.6881665249243496. min_data_in_leaf, val_score: 0.688167: 60%|###### | 3/5 [00:02<00:01, 1.14it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000967 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664957 valid's binary_logloss: 0.688576 Early stopping, best iteration is: [99] train's binary_logloss: 0.665189 valid's binary_logloss: 0.688543 min_data_in_leaf, val_score: 0.688167: 80%|######## | 4/5 [00:03<00:00, 1.09it/s][I 2020-09-27 04:53:59,775] Trial 66 finished with value: 0.6885432542311284 and parameters: {'min_child_samples': 5}. Best is trial 65 with value: 0.6881665249243496. min_data_in_leaf, val_score: 0.688167: 80%|######## | 4/5 [00:03<00:00, 1.09it/s][LightGBM] [Info] Number of positive: 46729, number of negative: 46297 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001196 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.502322 -> initscore=0.009288 [LightGBM] [Info] Start training from score 0.009288 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665534 valid's binary_logloss: 0.689168 Early stopping, best iteration is: [86] train's binary_logloss: 0.668451 valid's binary_logloss: 0.688864 min_data_in_leaf, val_score: 0.688167: 100%|##########| 5/5 [00:04<00:00, 1.07it/s][I 2020-09-27 04:54:00,741] Trial 67 finished with value: 0.6888639866159642 and parameters: {'min_child_samples': 50}. Best is trial 65 with value: 0.6881665249243496. min_data_in_leaf, val_score: 0.688167: 100%|##########| 5/5 [00:04<00:00, 1.08it/s] Fold : 3 [I 2020-09-27 04:54:00,831] A new study created in memory with name: no-name-2a0ad92a-db0c-44a0-a535-c4566ac6f3da feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010784 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.66381 valid's binary_logloss: 0.689338 Early stopping, best iteration is: [57] train's binary_logloss: 0.673532 valid's binary_logloss: 0.68897 feature_fraction, val_score: 0.688970: 14%|#4 | 1/7 [00:00<00:05, 1.01it/s][I 2020-09-27 04:54:01,833] Trial 0 finished with value: 0.6889700252608528 and parameters: {'feature_fraction': 0.7}. Best is trial 0 with value: 0.6889700252608528. feature_fraction, val_score: 0.688970: 14%|#4 | 1/7 [00:00<00:05, 1.01it/s][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001634 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.66267 valid's binary_logloss: 0.689828 Early stopping, best iteration is: [65] train's binary_logloss: 0.670714 valid's binary_logloss: 0.689441 feature_fraction, val_score: 0.688970: 29%|##8 | 2/7 [00:01<00:04, 1.04it/s][I 2020-09-27 04:54:02,732] Trial 1 finished with value: 0.6894408688583004 and parameters: {'feature_fraction': 1.0}. Best is trial 0 with value: 0.6889700252608528. feature_fraction, val_score: 0.688970: 29%|##8 | 2/7 [00:01<00:04, 1.04it/s][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000679 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.666188 valid's binary_logloss: 0.689964 [200] train's binary_logloss: 0.648164 valid's binary_logloss: 0.690421 Early stopping, best iteration is: [146] train's binary_logloss: 0.657493 valid's binary_logloss: 0.689838 feature_fraction, val_score: 0.688970: 43%|####2 | 3/7 [00:02<00:04, 1.00s/it][I 2020-09-27 04:54:03,817] Trial 2 finished with value: 0.6898376157358569 and parameters: {'feature_fraction': 0.4}. Best is trial 0 with value: 0.6889700252608528. feature_fraction, val_score: 0.688970: 43%|####2 | 3/7 [00:02<00:04, 1.00s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008176 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664448 valid's binary_logloss: 0.690083 Early stopping, best iteration is: [61] train's binary_logloss: 0.673008 valid's binary_logloss: 0.68947 feature_fraction, val_score: 0.688970: 57%|#####7 | 4/7 [00:03<00:02, 1.08it/s][I 2020-09-27 04:54:04,577] Trial 3 finished with value: 0.6894700918769894 and parameters: {'feature_fraction': 0.6}. Best is trial 0 with value: 0.6889700252608528. feature_fraction, val_score: 0.688970: 57%|#####7 | 4/7 [00:03<00:02, 1.08it/s][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001461 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.6633 valid's binary_logloss: 0.68927 [200] train's binary_logloss: 0.642614 valid's binary_logloss: 0.689045 Early stopping, best iteration is: [194] train's binary_logloss: 0.643867 valid's binary_logloss: 0.688894 feature_fraction, val_score: 0.688894: 71%|#######1 | 5/7 [00:05<00:02, 1.10s/it][I 2020-09-27 04:54:06,090] Trial 4 finished with value: 0.6888944409719155 and parameters: {'feature_fraction': 0.8}. Best is trial 4 with value: 0.6888944409719155. feature_fraction, val_score: 0.688894: 71%|#######1 | 5/7 [00:05<00:02, 1.10s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001711 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.662969 valid's binary_logloss: 0.689906 Early stopping, best iteration is: [58] train's binary_logloss: 0.672958 valid's binary_logloss: 0.689479 feature_fraction, val_score: 0.688894: 86%|########5 | 6/7 [00:06<00:01, 1.04s/it][I 2020-09-27 04:54:06,969] Trial 5 finished with value: 0.6894794839977714 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 4 with value: 0.6888944409719155. feature_fraction, val_score: 0.688894: 86%|########5 | 6/7 [00:06<00:01, 1.04s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001143 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664948 valid's binary_logloss: 0.689939 Early stopping, best iteration is: [71] train's binary_logloss: 0.671105 valid's binary_logloss: 0.689412 feature_fraction, val_score: 0.688894: 100%|##########| 7/7 [00:06<00:00, 1.02it/s][I 2020-09-27 04:54:07,829] Trial 6 finished with value: 0.689411673782528 and parameters: {'feature_fraction': 0.5}. Best is trial 4 with value: 0.6888944409719155. feature_fraction, val_score: 0.688894: 100%|##########| 7/7 [00:06<00:00, 1.00it/s] num_leaves, val_score: 0.688894: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016008 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.545996 valid's binary_logloss: 0.696346 Early stopping, best iteration is: [20] train's binary_logloss: 0.650465 valid's binary_logloss: 0.691113 num_leaves, val_score: 0.688894: 5%|5 | 1/20 [00:01<00:25, 1.32s/it][I 2020-09-27 04:54:09,164] Trial 7 finished with value: 0.6911130787415561 and parameters: {'num_leaves': 219}. Best is trial 7 with value: 0.6911130787415561. num_leaves, val_score: 0.688894: 5%|5 | 1/20 [00:01<00:25, 1.32s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001458 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.583507 valid's binary_logloss: 0.694216 Early stopping, best iteration is: [18] train's binary_logloss: 0.66415 valid's binary_logloss: 0.691315 num_leaves, val_score: 0.688894: 10%|# | 2/20 [00:02<00:23, 1.28s/it][I 2020-09-27 04:54:10,338] Trial 8 finished with value: 0.6913152010105598 and parameters: {'num_leaves': 152}. Best is trial 7 with value: 0.6911130787415561. num_leaves, val_score: 0.688894: 10%|# | 2/20 [00:02<00:23, 1.28s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001515 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.600147 valid's binary_logloss: 0.692792 Early stopping, best iteration is: [45] train's binary_logloss: 0.642663 valid's binary_logloss: 0.690464 num_leaves, val_score: 0.688894: 15%|#5 | 3/20 [00:04<00:23, 1.39s/it][I 2020-09-27 04:54:11,983] Trial 9 finished with value: 0.690463587315969 and parameters: {'num_leaves': 123}. Best is trial 9 with value: 0.690463587315969. num_leaves, val_score: 0.688894: 15%|#5 | 3/20 [00:04<00:23, 1.39s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001721 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.676573 valid's binary_logloss: 0.689403 Early stopping, best iteration is: [82] train's binary_logloss: 0.678598 valid's binary_logloss: 0.689178 num_leaves, val_score: 0.688894: 20%|## | 4/20 [00:06<00:25, 1.61s/it][I 2020-09-27 04:54:14,106] Trial 10 finished with value: 0.6891784496510537 and parameters: {'num_leaves': 15}. Best is trial 10 with value: 0.6891784496510537. num_leaves, val_score: 0.688894: 20%|## | 4/20 [00:06<00:25, 1.61s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012855 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689973 valid's binary_logloss: 0.691198 [200] train's binary_logloss: 0.688794 valid's binary_logloss: 0.690538 [300] train's binary_logloss: 0.68806 valid's binary_logloss: 0.690049 [400] train's binary_logloss: 0.687537 valid's binary_logloss: 0.689777 [500] train's binary_logloss: 0.687144 valid's binary_logloss: 0.689665 [600] train's binary_logloss: 0.686839 valid's binary_logloss: 0.689559 [700] train's binary_logloss: 0.686592 valid's binary_logloss: 0.689472 [800] train's binary_logloss: 0.686384 valid's binary_logloss: 0.689469 [900] train's binary_logloss: 0.686205 valid's binary_logloss: 0.689458 Early stopping, best iteration is: [877] train's binary_logloss: 0.686244 valid's binary_logloss: 0.68944 num_leaves, val_score: 0.688894: 25%|##5 | 5/20 [00:09<00:31, 2.13s/it][I 2020-09-27 04:54:17,448] Trial 11 finished with value: 0.689440488741354 and parameters: {'num_leaves': 2}. Best is trial 10 with value: 0.6891784496510537. num_leaves, val_score: 0.688894: 25%|##5 | 5/20 [00:09<00:31, 2.13s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001640 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.531631 valid's binary_logloss: 0.696171 Early stopping, best iteration is: [9] train's binary_logloss: 0.669502 valid's binary_logloss: 0.691204 num_leaves, val_score: 0.688894: 30%|### | 6/20 [00:11<00:27, 1.95s/it][I 2020-09-27 04:54:18,964] Trial 12 finished with value: 0.6912044508576668 and parameters: {'num_leaves': 252}. Best is trial 10 with value: 0.6891784496510537. num_leaves, val_score: 0.688894: 30%|### | 6/20 [00:11<00:27, 1.95s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001636 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.626032 valid's binary_logloss: 0.690485 Early stopping, best iteration is: [80] train's binary_logloss: 0.636438 valid's binary_logloss: 0.689817 num_leaves, val_score: 0.688894: 35%|###5 | 7/20 [00:12<00:22, 1.75s/it][I 2020-09-27 04:54:20,263] Trial 13 finished with value: 0.6898168565458592 and parameters: {'num_leaves': 83}. Best is trial 10 with value: 0.6891784496510537. num_leaves, val_score: 0.688894: 35%|###5 | 7/20 [00:12<00:22, 1.75s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010079 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.559692 valid's binary_logloss: 0.696329 Early stopping, best iteration is: [18] train's binary_logloss: 0.657719 valid's binary_logloss: 0.691223 num_leaves, val_score: 0.688894: 40%|#### | 8/20 [00:13<00:18, 1.58s/it][I 2020-09-27 04:54:21,443] Trial 14 finished with value: 0.6912227645026857 and parameters: {'num_leaves': 197}. Best is trial 10 with value: 0.6891784496510537. num_leaves, val_score: 0.688894: 40%|#### | 8/20 [00:13<00:18, 1.58s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001528 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.643665 valid's binary_logloss: 0.691255 Early stopping, best iteration is: [27] train's binary_logloss: 0.67467 valid's binary_logloss: 0.689948 num_leaves, val_score: 0.688894: 45%|####5 | 9/20 [00:14<00:14, 1.36s/it][I 2020-09-27 04:54:22,296] Trial 15 finished with value: 0.6899483024551792 and parameters: {'num_leaves': 57}. Best is trial 10 with value: 0.6891784496510537. num_leaves, val_score: 0.688894: 45%|####5 | 9/20 [00:14<00:14, 1.36s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001941 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.5757 valid's binary_logloss: 0.695653 Early stopping, best iteration is: [17] train's binary_logloss: 0.663745 valid's binary_logloss: 0.691071 num_leaves, val_score: 0.688894: 50%|##### | 10/20 [00:15<00:13, 1.32s/it][I 2020-09-27 04:54:23,534] Trial 16 finished with value: 0.6910709930215578 and parameters: {'num_leaves': 163}. Best is trial 10 with value: 0.6891784496510537. num_leaves, val_score: 0.688894: 50%|##### | 10/20 [00:15<00:13, 1.32s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008405 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.642489 valid's binary_logloss: 0.690471 Early stopping, best iteration is: [45] train's binary_logloss: 0.665549 valid's binary_logloss: 0.689944 num_leaves, val_score: 0.688894: 55%|#####5 | 11/20 [00:16<00:10, 1.20s/it][I 2020-09-27 04:54:24,436] Trial 17 finished with value: 0.6899435817261527 and parameters: {'num_leaves': 58}. Best is trial 10 with value: 0.6891784496510537. num_leaves, val_score: 0.688894: 55%|#####5 | 11/20 [00:16<00:10, 1.20s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002315 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.536859 valid's binary_logloss: 0.696258 Early stopping, best iteration is: [24] train's binary_logloss: 0.640234 valid's binary_logloss: 0.690993 num_leaves, val_score: 0.688894: 60%|###### | 12/20 [00:18<00:10, 1.32s/it][I 2020-09-27 04:54:26,057] Trial 18 finished with value: 0.6909927461103317 and parameters: {'num_leaves': 239}. Best is trial 10 with value: 0.6891784496510537. num_leaves, val_score: 0.688894: 60%|###### | 12/20 [00:18<00:10, 1.32s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001510 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.618426 valid's binary_logloss: 0.69242 Early stopping, best iteration is: [22] train's binary_logloss: 0.669782 valid's binary_logloss: 0.69036 num_leaves, val_score: 0.688894: 65%|######5 | 13/20 [00:19<00:08, 1.22s/it][I 2020-09-27 04:54:27,037] Trial 19 finished with value: 0.6903597022330396 and parameters: {'num_leaves': 95}. Best is trial 10 with value: 0.6891784496510537. num_leaves, val_score: 0.688894: 65%|######5 | 13/20 [00:19<00:08, 1.22s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.014205 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.568032 valid's binary_logloss: 0.694802 Early stopping, best iteration is: [20] train's binary_logloss: 0.656477 valid's binary_logloss: 0.690174 num_leaves, val_score: 0.688894: 70%|####### | 14/20 [00:20<00:07, 1.20s/it][I 2020-09-27 04:54:28,190] Trial 20 finished with value: 0.690174257788127 and parameters: {'num_leaves': 182}. Best is trial 10 with value: 0.6891784496510537. num_leaves, val_score: 0.688894: 70%|####### | 14/20 [00:20<00:07, 1.20s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.006271 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681177 valid's binary_logloss: 0.689414 [200] train's binary_logloss: 0.674276 valid's binary_logloss: 0.689006 Early stopping, best iteration is: [194] train's binary_logloss: 0.674612 valid's binary_logloss: 0.688884 num_leaves, val_score: 0.688884: 75%|#######5 | 15/20 [00:21<00:06, 1.24s/it][I 2020-09-27 04:54:29,535] Trial 21 finished with value: 0.6888838019748574 and parameters: {'num_leaves': 10}. Best is trial 21 with value: 0.6888838019748574. num_leaves, val_score: 0.688884: 75%|#######5 | 15/20 [00:21<00:06, 1.24s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001987 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663848 valid's binary_logloss: 0.689368 Early stopping, best iteration is: [90] train's binary_logloss: 0.666139 valid's binary_logloss: 0.689192 num_leaves, val_score: 0.688884: 80%|######## | 16/20 [00:22<00:04, 1.19s/it][I 2020-09-27 04:54:30,589] Trial 22 finished with value: 0.6891919324492329 and parameters: {'num_leaves': 30}. Best is trial 21 with value: 0.6888838019748574. num_leaves, val_score: 0.688884: 80%|######## | 16/20 [00:22<00:04, 1.19s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009391 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.605617 valid's binary_logloss: 0.693796 Early stopping, best iteration is: [22] train's binary_logloss: 0.666098 valid's binary_logloss: 0.690873 num_leaves, val_score: 0.688884: 85%|########5 | 17/20 [00:23<00:03, 1.13s/it][I 2020-09-27 04:54:31,594] Trial 23 finished with value: 0.6908732323148996 and parameters: {'num_leaves': 114}. Best is trial 21 with value: 0.6888838019748574. num_leaves, val_score: 0.688884: 85%|########5 | 17/20 [00:23<00:03, 1.13s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001645 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.655426 valid's binary_logloss: 0.689235 [200] train's binary_logloss: 0.628962 valid's binary_logloss: 0.691293 Early stopping, best iteration is: [106] train's binary_logloss: 0.653726 valid's binary_logloss: 0.689075 num_leaves, val_score: 0.688884: 90%|######### | 18/20 [00:24<00:02, 1.14s/it][I 2020-09-27 04:54:32,740] Trial 24 finished with value: 0.6890748065383784 and parameters: {'num_leaves': 41}. Best is trial 21 with value: 0.6888838019748574. num_leaves, val_score: 0.688884: 90%|######### | 18/20 [00:24<00:02, 1.14s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011385 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687114 valid's binary_logloss: 0.690103 [200] train's binary_logloss: 0.684519 valid's binary_logloss: 0.689685 [300] train's binary_logloss: 0.682489 valid's binary_logloss: 0.68949 [400] train's binary_logloss: 0.680652 valid's binary_logloss: 0.689206 [500] train's binary_logloss: 0.678959 valid's binary_logloss: 0.689203 Early stopping, best iteration is: [432] train's binary_logloss: 0.680085 valid's binary_logloss: 0.689077 num_leaves, val_score: 0.688884: 95%|#########5| 19/20 [00:26<00:01, 1.37s/it][I 2020-09-27 04:54:34,648] Trial 25 finished with value: 0.6890769719746892 and parameters: {'num_leaves': 4}. Best is trial 21 with value: 0.6888838019748574. num_leaves, val_score: 0.688884: 95%|#########5| 19/20 [00:26<00:01, 1.37s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002085 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.624994 valid's binary_logloss: 0.690549 Early stopping, best iteration is: [49] train's binary_logloss: 0.653531 valid's binary_logloss: 0.689478 num_leaves, val_score: 0.688884: 100%|##########| 20/20 [00:27<00:00, 1.30s/it][I 2020-09-27 04:54:35,793] Trial 26 finished with value: 0.6894778249561458 and parameters: {'num_leaves': 84}. Best is trial 21 with value: 0.6888838019748574. num_leaves, val_score: 0.688884: 100%|##########| 20/20 [00:27<00:00, 1.40s/it] bagging, val_score: 0.688884: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013914 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681106 valid's binary_logloss: 0.689431 [200] train's binary_logloss: 0.674446 valid's binary_logloss: 0.689212 [300] train's binary_logloss: 0.668203 valid's binary_logloss: 0.689716 Early stopping, best iteration is: [201] train's binary_logloss: 0.674354 valid's binary_logloss: 0.6892 bagging, val_score: 0.688884: 10%|# | 1/10 [00:01<00:11, 1.28s/it][I 2020-09-27 04:54:37,087] Trial 27 finished with value: 0.6891997209813286 and parameters: {'bagging_fraction': 0.5047216967290817, 'bagging_freq': 2}. Best is trial 27 with value: 0.6891997209813286. bagging, val_score: 0.688884: 10%|# | 1/10 [00:01<00:11, 1.28s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001692 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681088 valid's binary_logloss: 0.689391 [200] train's binary_logloss: 0.674086 valid's binary_logloss: 0.689322 Early stopping, best iteration is: [195] train's binary_logloss: 0.67442 valid's binary_logloss: 0.689229 bagging, val_score: 0.688884: 20%|## | 2/10 [00:02<00:10, 1.32s/it][I 2020-09-27 04:54:38,476] Trial 28 finished with value: 0.6892290270817055 and parameters: {'bagging_fraction': 0.9953814685875817, 'bagging_freq': 7}. Best is trial 27 with value: 0.6891997209813286. bagging, val_score: 0.688884: 20%|## | 2/10 [00:02<00:10, 1.32s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.014930 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681054 valid's binary_logloss: 0.689888 [200] train's binary_logloss: 0.674088 valid's binary_logloss: 0.689474 [300] train's binary_logloss: 0.667589 valid's binary_logloss: 0.689818 Early stopping, best iteration is: [201] train's binary_logloss: 0.674022 valid's binary_logloss: 0.689449 bagging, val_score: 0.688884: 30%|### | 3/10 [00:04<00:09, 1.34s/it][I 2020-09-27 04:54:39,878] Trial 29 finished with value: 0.6894491770687096 and parameters: {'bagging_fraction': 0.9657183733734178, 'bagging_freq': 7}. Best is trial 27 with value: 0.6891997209813286. bagging, val_score: 0.688884: 30%|### | 3/10 [00:04<00:09, 1.34s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001797 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681277 valid's binary_logloss: 0.689585 [200] train's binary_logloss: 0.674417 valid's binary_logloss: 0.689825 Early stopping, best iteration is: [161] train's binary_logloss: 0.676999 valid's binary_logloss: 0.689344 bagging, val_score: 0.688884: 40%|#### | 4/10 [00:05<00:07, 1.32s/it][I 2020-09-27 04:54:41,134] Trial 30 finished with value: 0.6893441175783354 and parameters: {'bagging_fraction': 0.4138320408319066, 'bagging_freq': 1}. Best is trial 27 with value: 0.6891997209813286. bagging, val_score: 0.688884: 40%|#### | 4/10 [00:05<00:07, 1.32s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008705 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681066 valid's binary_logloss: 0.688712 [200] train's binary_logloss: 0.674172 valid's binary_logloss: 0.688921 Early stopping, best iteration is: [121] train's binary_logloss: 0.679523 valid's binary_logloss: 0.688425 bagging, val_score: 0.688425: 50%|##### | 5/10 [00:06<00:06, 1.23s/it][I 2020-09-27 04:54:42,157] Trial 31 finished with value: 0.6884249102782825 and parameters: {'bagging_fraction': 0.7249814340678598, 'bagging_freq': 4}. Best is trial 31 with value: 0.6884249102782825. bagging, val_score: 0.688425: 50%|##### | 5/10 [00:06<00:06, 1.23s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001643 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681083 valid's binary_logloss: 0.689228 [200] train's binary_logloss: 0.674127 valid's binary_logloss: 0.688871 [300] train's binary_logloss: 0.667696 valid's binary_logloss: 0.688939 Early stopping, best iteration is: [247] train's binary_logloss: 0.671103 valid's binary_logloss: 0.688664 bagging, val_score: 0.688425: 60%|###### | 6/10 [00:07<00:05, 1.31s/it][I 2020-09-27 04:54:43,668] Trial 32 finished with value: 0.6886636187874674 and parameters: {'bagging_fraction': 0.7386084209140283, 'bagging_freq': 4}. Best is trial 31 with value: 0.6884249102782825. bagging, val_score: 0.688425: 60%|###### | 6/10 [00:07<00:05, 1.31s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001677 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681024 valid's binary_logloss: 0.689172 [200] train's binary_logloss: 0.67391 valid's binary_logloss: 0.689047 Early stopping, best iteration is: [125] train's binary_logloss: 0.67913 valid's binary_logloss: 0.688666 bagging, val_score: 0.688425: 70%|####### | 7/10 [00:08<00:03, 1.24s/it][I 2020-09-27 04:54:44,744] Trial 33 finished with value: 0.6886664091503993 and parameters: {'bagging_fraction': 0.765888231588719, 'bagging_freq': 4}. Best is trial 31 with value: 0.6884249102782825. bagging, val_score: 0.688425: 70%|####### | 7/10 [00:08<00:03, 1.24s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009276 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681076 valid's binary_logloss: 0.689153 [200] train's binary_logloss: 0.674066 valid's binary_logloss: 0.688436 [300] train's binary_logloss: 0.667548 valid's binary_logloss: 0.688896 Early stopping, best iteration is: [241] train's binary_logloss: 0.671442 valid's binary_logloss: 0.688118 bagging, val_score: 0.688118: 80%|######## | 8/10 [00:10<00:02, 1.31s/it][I 2020-09-27 04:54:46,218] Trial 34 finished with value: 0.6881176904104821 and parameters: {'bagging_fraction': 0.7616456433874508, 'bagging_freq': 4}. Best is trial 34 with value: 0.6881176904104821. bagging, val_score: 0.688118: 80%|######## | 8/10 [00:10<00:02, 1.31s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009038 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.680934 valid's binary_logloss: 0.688963 [200] train's binary_logloss: 0.674092 valid's binary_logloss: 0.68854 [300] train's binary_logloss: 0.667767 valid's binary_logloss: 0.688887 Early stopping, best iteration is: [208] train's binary_logloss: 0.673575 valid's binary_logloss: 0.688372 bagging, val_score: 0.688118: 90%|######### | 9/10 [00:11<00:01, 1.33s/it][I 2020-09-27 04:54:47,583] Trial 35 finished with value: 0.688371985997847 and parameters: {'bagging_fraction': 0.7538151922855485, 'bagging_freq': 4}. Best is trial 34 with value: 0.6881176904104821. bagging, val_score: 0.688118: 90%|######### | 9/10 [00:11<00:01, 1.33s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001608 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68108 valid's binary_logloss: 0.689706 [200] train's binary_logloss: 0.674027 valid's binary_logloss: 0.689423 Early stopping, best iteration is: [161] train's binary_logloss: 0.676626 valid's binary_logloss: 0.689208 bagging, val_score: 0.688118: 100%|##########| 10/10 [00:12<00:00, 1.29s/it][I 2020-09-27 04:54:48,782] Trial 36 finished with value: 0.6892084810261497 and parameters: {'bagging_fraction': 0.7387822613284701, 'bagging_freq': 4}. Best is trial 34 with value: 0.6881176904104821. bagging, val_score: 0.688118: 100%|##########| 10/10 [00:12<00:00, 1.30s/it] feature_fraction_stage2, val_score: 0.688118: 0%| | 0/6 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001509 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.680962 valid's binary_logloss: 0.689027 [200] train's binary_logloss: 0.673815 valid's binary_logloss: 0.689025 Early stopping, best iteration is: [122] train's binary_logloss: 0.679212 valid's binary_logloss: 0.688791 feature_fraction_stage2, val_score: 0.688118: 17%|#6 | 1/6 [00:01<00:05, 1.04s/it][I 2020-09-27 04:54:49,838] Trial 37 finished with value: 0.6887910798116761 and parameters: {'feature_fraction': 0.88}. Best is trial 37 with value: 0.6887910798116761. feature_fraction_stage2, val_score: 0.688118: 17%|#6 | 1/6 [00:01<00:05, 1.04s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008410 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681123 valid's binary_logloss: 0.688594 [200] train's binary_logloss: 0.674146 valid's binary_logloss: 0.688526 Early stopping, best iteration is: [128] train's binary_logloss: 0.679071 valid's binary_logloss: 0.688123 feature_fraction_stage2, val_score: 0.688118: 33%|###3 | 2/6 [00:02<00:04, 1.03s/it][I 2020-09-27 04:54:50,833] Trial 38 finished with value: 0.6881231868561525 and parameters: {'feature_fraction': 0.7200000000000001}. Best is trial 38 with value: 0.6881231868561525. feature_fraction_stage2, val_score: 0.688118: 33%|###3 | 2/6 [00:02<00:04, 1.03s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001586 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681076 valid's binary_logloss: 0.689153 [200] train's binary_logloss: 0.674066 valid's binary_logloss: 0.688436 [300] train's binary_logloss: 0.667548 valid's binary_logloss: 0.688896 Early stopping, best iteration is: [241] train's binary_logloss: 0.671442 valid's binary_logloss: 0.688118 feature_fraction_stage2, val_score: 0.688118: 50%|##### | 3/6 [00:03<00:03, 1.18s/it][I 2020-09-27 04:54:52,366] Trial 39 finished with value: 0.6881176904104821 and parameters: {'feature_fraction': 0.8160000000000001}. Best is trial 39 with value: 0.6881176904104821. feature_fraction_stage2, val_score: 0.688118: 50%|##### | 3/6 [00:03<00:03, 1.18s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011815 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681089 valid's binary_logloss: 0.689091 [200] train's binary_logloss: 0.674046 valid's binary_logloss: 0.68919 Early stopping, best iteration is: [125] train's binary_logloss: 0.679183 valid's binary_logloss: 0.688872 feature_fraction_stage2, val_score: 0.688118: 67%|######6 | 4/6 [00:04<00:02, 1.13s/it][I 2020-09-27 04:54:53,387] Trial 40 finished with value: 0.6888723605544023 and parameters: {'feature_fraction': 0.784}. Best is trial 39 with value: 0.6881176904104821. feature_fraction_stage2, val_score: 0.688118: 67%|######6 | 4/6 [00:04<00:02, 1.13s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008483 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681089 valid's binary_logloss: 0.689091 [200] train's binary_logloss: 0.674046 valid's binary_logloss: 0.68919 Early stopping, best iteration is: [125] train's binary_logloss: 0.679183 valid's binary_logloss: 0.688872 feature_fraction_stage2, val_score: 0.688118: 83%|########3 | 5/6 [00:05<00:01, 1.10s/it][I 2020-09-27 04:54:54,428] Trial 41 finished with value: 0.6888723605544023 and parameters: {'feature_fraction': 0.7520000000000001}. Best is trial 39 with value: 0.6881176904104821. feature_fraction_stage2, val_score: 0.688118: 83%|########3 | 5/6 [00:05<00:01, 1.10s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001564 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.680946 valid's binary_logloss: 0.689429 [200] train's binary_logloss: 0.67397 valid's binary_logloss: 0.689016 Early stopping, best iteration is: [190] train's binary_logloss: 0.674674 valid's binary_logloss: 0.688999 feature_fraction_stage2, val_score: 0.688118: 100%|##########| 6/6 [00:06<00:00, 1.17s/it][I 2020-09-27 04:54:55,767] Trial 42 finished with value: 0.6889992541114389 and parameters: {'feature_fraction': 0.8480000000000001}. Best is trial 39 with value: 0.6881176904104821. feature_fraction_stage2, val_score: 0.688118: 100%|##########| 6/6 [00:06<00:00, 1.16s/it] regularization_factors, val_score: 0.688118: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009845 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681108 valid's binary_logloss: 0.689185 [200] train's binary_logloss: 0.674108 valid's binary_logloss: 0.68916 Early stopping, best iteration is: [161] train's binary_logloss: 0.676632 valid's binary_logloss: 0.688793 regularization_factors, val_score: 0.688118: 5%|5 | 1/20 [00:01<00:22, 1.18s/it][I 2020-09-27 04:54:56,959] Trial 43 finished with value: 0.6887934098600209 and parameters: {'lambda_l1': 0.0384884910618084, 'lambda_l2': 1.204751116088014e-08}. Best is trial 43 with value: 0.6887934098600209. regularization_factors, val_score: 0.688118: 5%|5 | 1/20 [00:01<00:22, 1.18s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001587 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68186 valid's binary_logloss: 0.689223 [200] train's binary_logloss: 0.675896 valid's binary_logloss: 0.689073 Early stopping, best iteration is: [179] train's binary_logloss: 0.677134 valid's binary_logloss: 0.688878 regularization_factors, val_score: 0.688118: 10%|# | 2/20 [00:02<00:22, 1.22s/it][I 2020-09-27 04:54:58,283] Trial 44 finished with value: 0.6888779097684588 and parameters: {'lambda_l1': 1.6316340580802456e-08, 'lambda_l2': 9.98107290914818}. Best is trial 43 with value: 0.6887934098600209. regularization_factors, val_score: 0.688118: 10%|# | 2/20 [00:02<00:22, 1.22s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008407 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681077 valid's binary_logloss: 0.689153 [200] train's binary_logloss: 0.674066 valid's binary_logloss: 0.688436 [300] train's binary_logloss: 0.667549 valid's binary_logloss: 0.688896 Early stopping, best iteration is: [241] train's binary_logloss: 0.671443 valid's binary_logloss: 0.688118 regularization_factors, val_score: 0.688118: 15%|#5 | 3/20 [00:04<00:22, 1.31s/it][I 2020-09-27 04:54:59,812] Trial 45 finished with value: 0.688117656077511 and parameters: {'lambda_l1': 7.53282906565389e-08, 'lambda_l2': 0.000670356667923206}. Best is trial 45 with value: 0.688117656077511. regularization_factors, val_score: 0.688118: 15%|#5 | 3/20 [00:04<00:22, 1.31s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001515 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681077 valid's binary_logloss: 0.689153 [200] train's binary_logloss: 0.674066 valid's binary_logloss: 0.688436 [300] train's binary_logloss: 0.667548 valid's binary_logloss: 0.688896 Early stopping, best iteration is: [241] train's binary_logloss: 0.671443 valid's binary_logloss: 0.688118 regularization_factors, val_score: 0.688118: 20%|## | 4/20 [00:05<00:22, 1.39s/it][I 2020-09-27 04:55:01,389] Trial 46 finished with value: 0.6881176696895994 and parameters: {'lambda_l1': 2.804832879023663e-08, 'lambda_l2': 0.00040452126562955353}. Best is trial 45 with value: 0.688117656077511. regularization_factors, val_score: 0.688118: 20%|## | 4/20 [00:05<00:22, 1.39s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008716 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681077 valid's binary_logloss: 0.689153 [200] train's binary_logloss: 0.674066 valid's binary_logloss: 0.688436 [300] train's binary_logloss: 0.667549 valid's binary_logloss: 0.688896 Early stopping, best iteration is: [241] train's binary_logloss: 0.671443 valid's binary_logloss: 0.688118 regularization_factors, val_score: 0.688118: 25%|##5 | 5/20 [00:07<00:21, 1.43s/it][I 2020-09-27 04:55:02,896] Trial 47 finished with value: 0.6881176655585728 and parameters: {'lambda_l1': 1.5604063831480376e-08, 'lambda_l2': 0.0004850969258230225}. Best is trial 45 with value: 0.688117656077511. regularization_factors, val_score: 0.688118: 25%|##5 | 5/20 [00:07<00:21, 1.43s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001586 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681077 valid's binary_logloss: 0.689153 [200] train's binary_logloss: 0.674066 valid's binary_logloss: 0.688436 [300] train's binary_logloss: 0.667548 valid's binary_logloss: 0.688896 Early stopping, best iteration is: [241] train's binary_logloss: 0.671443 valid's binary_logloss: 0.688118 regularization_factors, val_score: 0.688118: 30%|### | 6/20 [00:08<00:20, 1.45s/it][I 2020-09-27 04:55:04,413] Trial 48 finished with value: 0.6881176673153139 and parameters: {'lambda_l1': 1.1804802009557963e-08, 'lambda_l2': 0.0004509060497221136}. Best is trial 45 with value: 0.688117656077511. regularization_factors, val_score: 0.688118: 30%|### | 6/20 [00:08<00:20, 1.45s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.015933 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681077 valid's binary_logloss: 0.689153 [200] train's binary_logloss: 0.674066 valid's binary_logloss: 0.688436 [300] train's binary_logloss: 0.667549 valid's binary_logloss: 0.688896 Early stopping, best iteration is: [241] train's binary_logloss: 0.671443 valid's binary_logloss: 0.688118 regularization_factors, val_score: 0.688118: 35%|###5 | 7/20 [00:10<00:19, 1.48s/it][I 2020-09-27 04:55:05,941] Trial 49 finished with value: 0.6881176626365213 and parameters: {'lambda_l1': 1.3636634898221364e-08, 'lambda_l2': 0.0005423517853601064}. Best is trial 45 with value: 0.688117656077511. regularization_factors, val_score: 0.688118: 35%|###5 | 7/20 [00:10<00:19, 1.48s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010107 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681077 valid's binary_logloss: 0.689153 [200] train's binary_logloss: 0.674066 valid's binary_logloss: 0.688436 [300] train's binary_logloss: 0.667549 valid's binary_logloss: 0.688896 Early stopping, best iteration is: [241] train's binary_logloss: 0.671443 valid's binary_logloss: 0.688118 regularization_factors, val_score: 0.688118: 40%|#### | 8/20 [00:11<00:17, 1.49s/it][I 2020-09-27 04:55:07,456] Trial 50 finished with value: 0.6881176473150824 and parameters: {'lambda_l1': 1.544452447154657e-08, 'lambda_l2': 0.000841525353074742}. Best is trial 50 with value: 0.6881176473150824. regularization_factors, val_score: 0.688118: 40%|#### | 8/20 [00:11<00:17, 1.49s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009452 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681077 valid's binary_logloss: 0.689153 [200] train's binary_logloss: 0.674066 valid's binary_logloss: 0.688436 [300] train's binary_logloss: 0.667549 valid's binary_logloss: 0.688896 Early stopping, best iteration is: [241] train's binary_logloss: 0.671443 valid's binary_logloss: 0.688118 regularization_factors, val_score: 0.688118: 45%|####5 | 9/20 [00:13<00:16, 1.50s/it][I 2020-09-27 04:55:08,986] Trial 51 finished with value: 0.6881176591565269 and parameters: {'lambda_l1': 1.1527007228477607e-08, 'lambda_l2': 0.0006103419407812519}. Best is trial 50 with value: 0.6881176473150824. regularization_factors, val_score: 0.688118: 45%|####5 | 9/20 [00:13<00:16, 1.50s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001914 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681077 valid's binary_logloss: 0.689153 [200] train's binary_logloss: 0.674066 valid's binary_logloss: 0.688436 [300] train's binary_logloss: 0.667549 valid's binary_logloss: 0.688896 Early stopping, best iteration is: [241] train's binary_logloss: 0.671443 valid's binary_logloss: 0.688118 regularization_factors, val_score: 0.688118: 50%|##### | 10/20 [00:14<00:15, 1.52s/it][I 2020-09-27 04:55:10,538] Trial 52 finished with value: 0.6881176497206014 and parameters: {'lambda_l1': 1.0151302543150935e-08, 'lambda_l2': 0.000794572846016082}. Best is trial 50 with value: 0.6881176473150824. regularization_factors, val_score: 0.688118: 50%|##### | 10/20 [00:14<00:15, 1.52s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008731 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681077 valid's binary_logloss: 0.689153 [200] train's binary_logloss: 0.674066 valid's binary_logloss: 0.688436 [300] train's binary_logloss: 0.667549 valid's binary_logloss: 0.688896 Early stopping, best iteration is: [241] train's binary_logloss: 0.671443 valid's binary_logloss: 0.688118 regularization_factors, val_score: 0.688118: 55%|#####5 | 11/20 [00:16<00:13, 1.52s/it][I 2020-09-27 04:55:12,058] Trial 53 finished with value: 0.6881176463901173 and parameters: {'lambda_l1': 1.6743100485936676e-08, 'lambda_l2': 0.0008594795917381767}. Best is trial 53 with value: 0.6881176463901173. regularization_factors, val_score: 0.688118: 55%|#####5 | 11/20 [00:16<00:13, 1.52s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009990 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681077 valid's binary_logloss: 0.689152 [200] train's binary_logloss: 0.674067 valid's binary_logloss: 0.688436 [300] train's binary_logloss: 0.667549 valid's binary_logloss: 0.688896 Early stopping, best iteration is: [241] train's binary_logloss: 0.671443 valid's binary_logloss: 0.688118 regularization_factors, val_score: 0.688118: 60%|###### | 12/20 [00:17<00:12, 1.52s/it][I 2020-09-27 04:55:13,591] Trial 54 finished with value: 0.6881176183828548 and parameters: {'lambda_l1': 1.3182804260599829e-08, 'lambda_l2': 0.0014065609790939724}. Best is trial 54 with value: 0.6881176183828548. regularization_factors, val_score: 0.688118: 60%|###### | 12/20 [00:17<00:12, 1.52s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001677 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681004 valid's binary_logloss: 0.689046 [200] train's binary_logloss: 0.674082 valid's binary_logloss: 0.688629 [300] train's binary_logloss: 0.667629 valid's binary_logloss: 0.688981 Early stopping, best iteration is: [208] train's binary_logloss: 0.673549 valid's binary_logloss: 0.68846 regularization_factors, val_score: 0.688118: 65%|######5 | 13/20 [00:19<00:10, 1.49s/it][I 2020-09-27 04:55:15,021] Trial 55 finished with value: 0.6884597615807989 and parameters: {'lambda_l1': 1.1894995962263343e-08, 'lambda_l2': 0.007778232962106016}. Best is trial 54 with value: 0.6881176183828548. regularization_factors, val_score: 0.688118: 65%|######5 | 13/20 [00:19<00:10, 1.49s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001586 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681005 valid's binary_logloss: 0.689046 [200] train's binary_logloss: 0.674033 valid's binary_logloss: 0.688993 Early stopping, best iteration is: [127] train's binary_logloss: 0.678994 valid's binary_logloss: 0.688676 regularization_factors, val_score: 0.688118: 70%|####### | 14/20 [00:20<00:08, 1.37s/it][I 2020-09-27 04:55:16,097] Trial 56 finished with value: 0.6886762297484497 and parameters: {'lambda_l1': 1.041076968285473e-06, 'lambda_l2': 0.011119399639455865}. Best is trial 54 with value: 0.6881176183828548. regularization_factors, val_score: 0.688118: 70%|####### | 14/20 [00:20<00:08, 1.37s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001534 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681077 valid's binary_logloss: 0.689152 [200] train's binary_logloss: 0.67413 valid's binary_logloss: 0.68893 Early stopping, best iteration is: [128] train's binary_logloss: 0.67903 valid's binary_logloss: 0.688711 regularization_factors, val_score: 0.688118: 75%|#######5 | 15/20 [00:21<00:06, 1.27s/it][I 2020-09-27 04:55:17,151] Trial 57 finished with value: 0.6887114735383899 and parameters: {'lambda_l1': 2.060141563353616e-07, 'lambda_l2': 0.004227573539355788}. Best is trial 54 with value: 0.6881176183828548. regularization_factors, val_score: 0.688118: 75%|#######5 | 15/20 [00:21<00:06, 1.27s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001568 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681076 valid's binary_logloss: 0.689153 [200] train's binary_logloss: 0.674066 valid's binary_logloss: 0.688436 [300] train's binary_logloss: 0.667548 valid's binary_logloss: 0.688896 Early stopping, best iteration is: [241] train's binary_logloss: 0.671442 valid's binary_logloss: 0.688118 regularization_factors, val_score: 0.688118: 80%|######## | 16/20 [00:22<00:05, 1.36s/it][I 2020-09-27 04:55:18,723] Trial 58 finished with value: 0.6881176889342293 and parameters: {'lambda_l1': 1.1056346896610525e-06, 'lambda_l2': 2.7165035646251157e-05}. Best is trial 54 with value: 0.6881176183828548. regularization_factors, val_score: 0.688118: 80%|######## | 16/20 [00:22<00:05, 1.36s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001614 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681076 valid's binary_logloss: 0.689153 [200] train's binary_logloss: 0.674066 valid's binary_logloss: 0.688436 [300] train's binary_logloss: 0.667548 valid's binary_logloss: 0.688896 Early stopping, best iteration is: [241] train's binary_logloss: 0.671442 valid's binary_logloss: 0.688118 regularization_factors, val_score: 0.688118: 85%|########5 | 17/20 [00:24<00:04, 1.42s/it][I 2020-09-27 04:55:20,289] Trial 59 finished with value: 0.6881176899381117 and parameters: {'lambda_l1': 1.4591391563330158e-07, 'lambda_l2': 9.037290952263432e-06}. Best is trial 54 with value: 0.6881176183828548. regularization_factors, val_score: 0.688118: 85%|########5 | 17/20 [00:24<00:04, 1.42s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008880 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681099 valid's binary_logloss: 0.689028 [200] train's binary_logloss: 0.674171 valid's binary_logloss: 0.688687 [300] train's binary_logloss: 0.667693 valid's binary_logloss: 0.689247 Early stopping, best iteration is: [208] train's binary_logloss: 0.67365 valid's binary_logloss: 0.688536 regularization_factors, val_score: 0.688118: 90%|######### | 18/20 [00:25<00:02, 1.43s/it][I 2020-09-27 04:55:21,723] Trial 60 finished with value: 0.6885355908433756 and parameters: {'lambda_l1': 1.3062472410110885e-08, 'lambda_l2': 0.007173981015512951}. Best is trial 54 with value: 0.6881176183828548. regularization_factors, val_score: 0.688118: 90%|######### | 18/20 [00:25<00:02, 1.43s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001684 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681077 valid's binary_logloss: 0.689153 [200] train's binary_logloss: 0.674066 valid's binary_logloss: 0.688436 [300] train's binary_logloss: 0.667548 valid's binary_logloss: 0.688896 Early stopping, best iteration is: [241] train's binary_logloss: 0.671443 valid's binary_logloss: 0.688118 regularization_factors, val_score: 0.688118: 95%|#########5| 19/20 [00:27<00:01, 1.47s/it][I 2020-09-27 04:55:23,279] Trial 61 finished with value: 0.6881176692604842 and parameters: {'lambda_l1': 1.0159772737706575e-08, 'lambda_l2': 0.0004128540970680604}. Best is trial 54 with value: 0.6881176183828548. regularization_factors, val_score: 0.688118: 95%|#########5| 19/20 [00:27<00:01, 1.47s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001565 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681077 valid's binary_logloss: 0.689153 [200] train's binary_logloss: 0.674067 valid's binary_logloss: 0.688436 [300] train's binary_logloss: 0.667549 valid's binary_logloss: 0.688896 Early stopping, best iteration is: [241] train's binary_logloss: 0.671443 valid's binary_logloss: 0.688118 regularization_factors, val_score: 0.688118: 100%|##########| 20/20 [00:29<00:00, 1.48s/it][I 2020-09-27 04:55:24,804] Trial 62 finished with value: 0.6881176378261832 and parameters: {'lambda_l1': 1.2256608662872115e-08, 'lambda_l2': 0.0010266987268648666}. Best is trial 54 with value: 0.6881176183828548. regularization_factors, val_score: 0.688118: 100%|##########| 20/20 [00:29<00:00, 1.45s/it] min_data_in_leaf, val_score: 0.688118: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001754 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681082 valid's binary_logloss: 0.68904 [200] train's binary_logloss: 0.674109 valid's binary_logloss: 0.689221 Early stopping, best iteration is: [127] train's binary_logloss: 0.679064 valid's binary_logloss: 0.688843 min_data_in_leaf, val_score: 0.688118: 20%|## | 1/5 [00:01<00:04, 1.11s/it][I 2020-09-27 04:55:25,927] Trial 63 finished with value: 0.6888427522145505 and parameters: {'min_child_samples': 25}. Best is trial 63 with value: 0.6888427522145505. min_data_in_leaf, val_score: 0.688118: 20%|## | 1/5 [00:01<00:04, 1.11s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001638 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68099 valid's binary_logloss: 0.689568 [200] train's binary_logloss: 0.67383 valid's binary_logloss: 0.68945 Early stopping, best iteration is: [139] train's binary_logloss: 0.678037 valid's binary_logloss: 0.689261 min_data_in_leaf, val_score: 0.688118: 40%|#### | 2/5 [00:02<00:03, 1.16s/it][I 2020-09-27 04:55:27,218] Trial 64 finished with value: 0.6892610898091048 and parameters: {'min_child_samples': 5}. Best is trial 63 with value: 0.6888427522145505. min_data_in_leaf, val_score: 0.688118: 40%|#### | 2/5 [00:02<00:03, 1.16s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002000 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681202 valid's binary_logloss: 0.689481 [200] train's binary_logloss: 0.674521 valid's binary_logloss: 0.689204 Early stopping, best iteration is: [133] train's binary_logloss: 0.678939 valid's binary_logloss: 0.688898 min_data_in_leaf, val_score: 0.688118: 60%|###### | 3/5 [00:03<00:02, 1.18s/it][I 2020-09-27 04:55:28,425] Trial 65 finished with value: 0.688897822488096 and parameters: {'min_child_samples': 50}. Best is trial 63 with value: 0.6888427522145505. min_data_in_leaf, val_score: 0.688118: 60%|###### | 3/5 [00:03<00:02, 1.18s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001549 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681041 valid's binary_logloss: 0.68912 [200] train's binary_logloss: 0.673932 valid's binary_logloss: 0.689165 [300] train's binary_logloss: 0.667472 valid's binary_logloss: 0.689403 Early stopping, best iteration is: [223] train's binary_logloss: 0.672471 valid's binary_logloss: 0.688731 min_data_in_leaf, val_score: 0.688118: 80%|######## | 4/5 [00:05<00:01, 1.28s/it][I 2020-09-27 04:55:29,951] Trial 66 finished with value: 0.6887314309367388 and parameters: {'min_child_samples': 10}. Best is trial 66 with value: 0.6887314309367388. min_data_in_leaf, val_score: 0.688118: 80%|######## | 4/5 [00:05<00:01, 1.28s/it][LightGBM] [Info] Number of positive: 46662, number of negative: 46364 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001500 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501602 -> initscore=0.006407 [LightGBM] [Info] Start training from score 0.006407 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.681418 valid's binary_logloss: 0.688989 [200] train's binary_logloss: 0.675177 valid's binary_logloss: 0.688794 [300] train's binary_logloss: 0.669107 valid's binary_logloss: 0.688977 Early stopping, best iteration is: [242] train's binary_logloss: 0.672665 valid's binary_logloss: 0.688301 min_data_in_leaf, val_score: 0.688118: 100%|##########| 5/5 [00:06<00:00, 1.38s/it][I 2020-09-27 04:55:31,562] Trial 67 finished with value: 0.6883012831237107 and parameters: {'min_child_samples': 100}. Best is trial 67 with value: 0.6883012831237107. min_data_in_leaf, val_score: 0.688118: 100%|##########| 5/5 [00:06<00:00, 1.35s/it] Fold : 4 [I 2020-09-27 04:55:31,677] A new study created in memory with name: no-name-9f8068ee-5ed7-40e8-b31b-fe76cdedea06 feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008972 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664047 valid's binary_logloss: 0.690512 Early stopping, best iteration is: [55] train's binary_logloss: 0.674377 valid's binary_logloss: 0.689789 feature_fraction, val_score: 0.689789: 14%|#4 | 1/7 [00:00<00:05, 1.05it/s][I 2020-09-27 04:55:32,636] Trial 0 finished with value: 0.6897889878374779 and parameters: {'feature_fraction': 0.6}. Best is trial 0 with value: 0.6897889878374779. feature_fraction, val_score: 0.689789: 14%|#4 | 1/7 [00:00<00:05, 1.05it/s][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000871 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.666186 valid's binary_logloss: 0.690817 Early stopping, best iteration is: [41] train's binary_logloss: 0.679056 valid's binary_logloss: 0.69044 feature_fraction, val_score: 0.689789: 29%|##8 | 2/7 [00:01<00:04, 1.14it/s][I 2020-09-27 04:55:33,334] Trial 1 finished with value: 0.6904400471655459 and parameters: {'feature_fraction': 0.4}. Best is trial 0 with value: 0.6897889878374779. feature_fraction, val_score: 0.689789: 29%|##8 | 2/7 [00:01<00:04, 1.14it/s][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000896 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.66495 valid's binary_logloss: 0.68987 Early stopping, best iteration is: [42] train's binary_logloss: 0.678016 valid's binary_logloss: 0.689555 feature_fraction, val_score: 0.689555: 43%|####2 | 3/7 [00:02<00:03, 1.19it/s][I 2020-09-27 04:55:34,092] Trial 2 finished with value: 0.6895553636305107 and parameters: {'feature_fraction': 0.5}. Best is trial 2 with value: 0.6895553636305107. feature_fraction, val_score: 0.689555: 43%|####2 | 3/7 [00:02<00:03, 1.19it/s][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.014383 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663818 valid's binary_logloss: 0.690283 Early stopping, best iteration is: [60] train's binary_logloss: 0.672948 valid's binary_logloss: 0.689614 feature_fraction, val_score: 0.689555: 57%|#####7 | 4/7 [00:03<00:02, 1.20it/s][I 2020-09-27 04:55:34,910] Trial 3 finished with value: 0.689613792253809 and parameters: {'feature_fraction': 0.7}. Best is trial 2 with value: 0.6895553636305107. feature_fraction, val_score: 0.689555: 57%|#####7 | 4/7 [00:03<00:02, 1.20it/s][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.007348 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663516 valid's binary_logloss: 0.691084 Early stopping, best iteration is: [56] train's binary_logloss: 0.673734 valid's binary_logloss: 0.690379 feature_fraction, val_score: 0.689555: 71%|#######1 | 5/7 [00:04<00:01, 1.18it/s][I 2020-09-27 04:55:35,802] Trial 4 finished with value: 0.6903789869460358 and parameters: {'feature_fraction': 0.8}. Best is trial 2 with value: 0.6895553636305107. feature_fraction, val_score: 0.689555: 71%|#######1 | 5/7 [00:04<00:01, 1.18it/s][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001795 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.662299 valid's binary_logloss: 0.689516 Early stopping, best iteration is: [43] train's binary_logloss: 0.676528 valid's binary_logloss: 0.689224 feature_fraction, val_score: 0.689224: 86%|########5 | 6/7 [00:05<00:00, 1.15it/s][I 2020-09-27 04:55:36,710] Trial 5 finished with value: 0.689223941248766 and parameters: {'feature_fraction': 1.0}. Best is trial 5 with value: 0.689223941248766. feature_fraction, val_score: 0.689224: 86%|########5 | 6/7 [00:05<00:00, 1.15it/s][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.005320 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663036 valid's binary_logloss: 0.690663 Early stopping, best iteration is: [45] train's binary_logloss: 0.676494 valid's binary_logloss: 0.689393 feature_fraction, val_score: 0.689224: 100%|##########| 7/7 [00:06<00:00, 1.10it/s][I 2020-09-27 04:55:37,710] Trial 6 finished with value: 0.6893925766913008 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 5 with value: 0.689223941248766. feature_fraction, val_score: 0.689224: 100%|##########| 7/7 [00:06<00:00, 1.16it/s] num_leaves, val_score: 0.689224: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001743 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.55132 valid's binary_logloss: 0.697443 Early stopping, best iteration is: [11] train's binary_logloss: 0.668724 valid's binary_logloss: 0.691123 num_leaves, val_score: 0.689224: 5%|5 | 1/20 [00:01<00:26, 1.38s/it][I 2020-09-27 04:55:39,105] Trial 7 finished with value: 0.6911225737618327 and parameters: {'num_leaves': 200}. Best is trial 7 with value: 0.6911225737618327. num_leaves, val_score: 0.689224: 5%|5 | 1/20 [00:01<00:26, 1.38s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001580 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.571031 valid's binary_logloss: 0.698616 Early stopping, best iteration is: [13] train's binary_logloss: 0.669076 valid's binary_logloss: 0.690967 num_leaves, val_score: 0.689224: 10%|# | 2/20 [00:02<00:25, 1.39s/it][I 2020-09-27 04:55:40,522] Trial 8 finished with value: 0.690966972517542 and parameters: {'num_leaves': 165}. Best is trial 8 with value: 0.690966972517542. num_leaves, val_score: 0.689224: 10%|# | 2/20 [00:02<00:25, 1.39s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002611 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.608797 valid's binary_logloss: 0.692752 Early stopping, best iteration is: [19] train's binary_logloss: 0.670075 valid's binary_logloss: 0.690291 num_leaves, val_score: 0.689224: 15%|#5 | 3/20 [00:04<00:23, 1.36s/it][I 2020-09-27 04:55:41,828] Trial 9 finished with value: 0.690291129092192 and parameters: {'num_leaves': 104}. Best is trial 9 with value: 0.690291129092192. num_leaves, val_score: 0.689224: 15%|#5 | 3/20 [00:04<00:23, 1.36s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001684 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687085 valid's binary_logloss: 0.689639 [200] train's binary_logloss: 0.684519 valid's binary_logloss: 0.689387 [300] train's binary_logloss: 0.682492 valid's binary_logloss: 0.689351 Early stopping, best iteration is: [285] train's binary_logloss: 0.682778 valid's binary_logloss: 0.689263 num_leaves, val_score: 0.689224: 20%|## | 4/20 [00:05<00:23, 1.47s/it][I 2020-09-27 04:55:43,548] Trial 10 finished with value: 0.6892629508973852 and parameters: {'num_leaves': 4}. Best is trial 10 with value: 0.6892629508973852. num_leaves, val_score: 0.689224: 20%|## | 4/20 [00:05<00:23, 1.47s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002867 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687085 valid's binary_logloss: 0.689639 [200] train's binary_logloss: 0.684519 valid's binary_logloss: 0.689387 [300] train's binary_logloss: 0.682492 valid's binary_logloss: 0.689351 Early stopping, best iteration is: [285] train's binary_logloss: 0.682778 valid's binary_logloss: 0.689263 num_leaves, val_score: 0.689224: 25%|##5 | 5/20 [00:07<00:24, 1.62s/it][I 2020-09-27 04:55:45,500] Trial 11 finished with value: 0.6892629508973852 and parameters: {'num_leaves': 4}. Best is trial 10 with value: 0.6892629508973852. num_leaves, val_score: 0.689224: 25%|##5 | 5/20 [00:07<00:24, 1.62s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.003844 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.670769 valid's binary_logloss: 0.690136 Early stopping, best iteration is: [57] train's binary_logloss: 0.67811 valid's binary_logloss: 0.689439 num_leaves, val_score: 0.689224: 30%|### | 6/20 [00:08<00:20, 1.46s/it][I 2020-09-27 04:55:46,606] Trial 12 finished with value: 0.6894394773112021 and parameters: {'num_leaves': 21}. Best is trial 10 with value: 0.6892629508973852. num_leaves, val_score: 0.689224: 30%|### | 6/20 [00:08<00:20, 1.46s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002745 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.632184 valid's binary_logloss: 0.690411 Early stopping, best iteration is: [41] train's binary_logloss: 0.662296 valid's binary_logloss: 0.689656 num_leaves, val_score: 0.689224: 35%|###5 | 7/20 [00:09<00:17, 1.35s/it][I 2020-09-27 04:55:47,703] Trial 13 finished with value: 0.6896563634195803 and parameters: {'num_leaves': 71}. Best is trial 10 with value: 0.6892629508973852. num_leaves, val_score: 0.689224: 35%|###5 | 7/20 [00:09<00:17, 1.35s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001681 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.528217 valid's binary_logloss: 0.698531 Early stopping, best iteration is: [10] train's binary_logloss: 0.667084 valid's binary_logloss: 0.692043 num_leaves, val_score: 0.689224: 40%|#### | 8/20 [00:11<00:16, 1.39s/it][I 2020-09-27 04:55:49,169] Trial 14 finished with value: 0.6920431771973239 and parameters: {'num_leaves': 242}. Best is trial 10 with value: 0.6892629508973852. num_leaves, val_score: 0.689224: 40%|#### | 8/20 [00:11<00:16, 1.39s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001737 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.647395 valid's binary_logloss: 0.692015 Early stopping, best iteration is: [18] train's binary_logloss: 0.680717 valid's binary_logloss: 0.689923 num_leaves, val_score: 0.689224: 45%|####5 | 9/20 [00:12<00:13, 1.19s/it][I 2020-09-27 04:55:49,908] Trial 15 finished with value: 0.6899230440367558 and parameters: {'num_leaves': 50}. Best is trial 10 with value: 0.6892629508973852. num_leaves, val_score: 0.689224: 45%|####5 | 9/20 [00:12<00:13, 1.19s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001534 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.599344 valid's binary_logloss: 0.693711 Early stopping, best iteration is: [20] train's binary_logloss: 0.666348 valid's binary_logloss: 0.689963 num_leaves, val_score: 0.689224: 50%|##### | 10/20 [00:13<00:11, 1.16s/it][I 2020-09-27 04:55:50,983] Trial 16 finished with value: 0.6899633518471027 and parameters: {'num_leaves': 119}. Best is trial 10 with value: 0.6892629508973852. num_leaves, val_score: 0.689224: 50%|##### | 10/20 [00:13<00:11, 1.16s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001595 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688413 valid's binary_logloss: 0.689959 [200] train's binary_logloss: 0.686555 valid's binary_logloss: 0.689381 [300] train's binary_logloss: 0.685253 valid's binary_logloss: 0.688961 [400] train's binary_logloss: 0.684142 valid's binary_logloss: 0.689082 Early stopping, best iteration is: [301] train's binary_logloss: 0.685239 valid's binary_logloss: 0.688951 num_leaves, val_score: 0.688951: 55%|#####5 | 11/20 [00:14<00:11, 1.24s/it][I 2020-09-27 04:55:52,401] Trial 17 finished with value: 0.6889511525087615 and parameters: {'num_leaves': 3}. Best is trial 17 with value: 0.6889511525087615. num_leaves, val_score: 0.688951: 55%|#####5 | 11/20 [00:14<00:11, 1.24s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002030 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.571031 valid's binary_logloss: 0.698616 Early stopping, best iteration is: [13] train's binary_logloss: 0.669076 valid's binary_logloss: 0.690967 num_leaves, val_score: 0.688951: 60%|###### | 12/20 [00:16<00:10, 1.27s/it][I 2020-09-27 04:55:53,737] Trial 18 finished with value: 0.690966972517542 and parameters: {'num_leaves': 165}. Best is trial 17 with value: 0.6889511525087615. num_leaves, val_score: 0.688951: 60%|###### | 12/20 [00:16<00:10, 1.27s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001718 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.628315 valid's binary_logloss: 0.692025 Early stopping, best iteration is: [25] train's binary_logloss: 0.670941 valid's binary_logloss: 0.689909 num_leaves, val_score: 0.688951: 65%|######5 | 13/20 [00:16<00:08, 1.17s/it][I 2020-09-27 04:55:54,692] Trial 19 finished with value: 0.6899089284634762 and parameters: {'num_leaves': 76}. Best is trial 17 with value: 0.6889511525087615. num_leaves, val_score: 0.688951: 65%|######5 | 13/20 [00:16<00:08, 1.17s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001740 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656752 valid's binary_logloss: 0.690382 Early stopping, best iteration is: [52] train's binary_logloss: 0.670809 valid's binary_logloss: 0.689428 num_leaves, val_score: 0.688951: 70%|####### | 14/20 [00:17<00:06, 1.08s/it][I 2020-09-27 04:55:55,563] Trial 20 finished with value: 0.6894277899501294 and parameters: {'num_leaves': 38}. Best is trial 17 with value: 0.6889511525087615. num_leaves, val_score: 0.688951: 70%|####### | 14/20 [00:17<00:06, 1.08s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001722 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687085 valid's binary_logloss: 0.689639 [200] train's binary_logloss: 0.684519 valid's binary_logloss: 0.689387 [300] train's binary_logloss: 0.682492 valid's binary_logloss: 0.689351 Early stopping, best iteration is: [285] train's binary_logloss: 0.682778 valid's binary_logloss: 0.689263 num_leaves, val_score: 0.688951: 75%|#######5 | 15/20 [00:19<00:05, 1.17s/it][I 2020-09-27 04:55:56,939] Trial 21 finished with value: 0.6892629508973853 and parameters: {'num_leaves': 4}. Best is trial 17 with value: 0.6889511525087615. num_leaves, val_score: 0.688951: 75%|#######5 | 15/20 [00:19<00:05, 1.17s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.005007 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687085 valid's binary_logloss: 0.689639 [200] train's binary_logloss: 0.684519 valid's binary_logloss: 0.689387 [300] train's binary_logloss: 0.682492 valid's binary_logloss: 0.689351 Early stopping, best iteration is: [285] train's binary_logloss: 0.682778 valid's binary_logloss: 0.689263 num_leaves, val_score: 0.688951: 80%|######## | 16/20 [00:20<00:04, 1.25s/it][I 2020-09-27 04:55:58,373] Trial 22 finished with value: 0.6892629508973852 and parameters: {'num_leaves': 4}. Best is trial 17 with value: 0.6889511525087615. num_leaves, val_score: 0.688951: 80%|######## | 16/20 [00:20<00:04, 1.25s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001737 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.634712 valid's binary_logloss: 0.690552 Early stopping, best iteration is: [28] train's binary_logloss: 0.671201 valid's binary_logloss: 0.689554 num_leaves, val_score: 0.688951: 85%|########5 | 17/20 [00:21<00:03, 1.17s/it][I 2020-09-27 04:55:59,359] Trial 23 finished with value: 0.6895543062404332 and parameters: {'num_leaves': 67}. Best is trial 17 with value: 0.6889511525087615. num_leaves, val_score: 0.688951: 85%|########5 | 17/20 [00:21<00:03, 1.17s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001626 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.660898 valid's binary_logloss: 0.690623 Early stopping, best iteration is: [44] train's binary_logloss: 0.675571 valid's binary_logloss: 0.689922 num_leaves, val_score: 0.688951: 90%|######### | 18/20 [00:22<00:02, 1.09s/it][I 2020-09-27 04:56:00,277] Trial 24 finished with value: 0.6899216528961626 and parameters: {'num_leaves': 33}. Best is trial 17 with value: 0.6889511525087615. num_leaves, val_score: 0.688951: 90%|######### | 18/20 [00:22<00:02, 1.09s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002059 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.614335 valid's binary_logloss: 0.691319 Early stopping, best iteration is: [19] train's binary_logloss: 0.67155 valid's binary_logloss: 0.689698 num_leaves, val_score: 0.688951: 95%|#########5| 19/20 [00:23<00:01, 1.08s/it][I 2020-09-27 04:56:01,311] Trial 25 finished with value: 0.6896983317380696 and parameters: {'num_leaves': 96}. Best is trial 17 with value: 0.6889511525087615. num_leaves, val_score: 0.688951: 95%|#########5| 19/20 [00:23<00:01, 1.08s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001688 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.58378 valid's binary_logloss: 0.69401 Early stopping, best iteration is: [20] train's binary_logloss: 0.662024 valid's binary_logloss: 0.690131 num_leaves, val_score: 0.688951: 100%|##########| 20/20 [00:24<00:00, 1.11s/it][I 2020-09-27 04:56:02,500] Trial 26 finished with value: 0.6901310985346112 and parameters: {'num_leaves': 143}. Best is trial 17 with value: 0.6889511525087615. num_leaves, val_score: 0.688951: 100%|##########| 20/20 [00:24<00:00, 1.24s/it] bagging, val_score: 0.688951: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013453 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688279 valid's binary_logloss: 0.689921 [200] train's binary_logloss: 0.686342 valid's binary_logloss: 0.68936 [300] train's binary_logloss: 0.685002 valid's binary_logloss: 0.688756 [400] train's binary_logloss: 0.683806 valid's binary_logloss: 0.688689 Early stopping, best iteration is: [358] train's binary_logloss: 0.684279 valid's binary_logloss: 0.688644 bagging, val_score: 0.688644: 10%|# | 1/10 [00:01<00:16, 1.81s/it][I 2020-09-27 04:56:04,317] Trial 27 finished with value: 0.6886437067357495 and parameters: {'bagging_fraction': 0.8690258816156133, 'bagging_freq': 3}. Best is trial 27 with value: 0.6886437067357495. bagging, val_score: 0.688644: 10%|# | 1/10 [00:01<00:16, 1.81s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001668 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688299 valid's binary_logloss: 0.689981 [200] train's binary_logloss: 0.686384 valid's binary_logloss: 0.689407 [300] train's binary_logloss: 0.684946 valid's binary_logloss: 0.689175 Early stopping, best iteration is: [276] train's binary_logloss: 0.685288 valid's binary_logloss: 0.689089 bagging, val_score: 0.688644: 20%|## | 2/10 [00:03<00:13, 1.73s/it][I 2020-09-27 04:56:05,876] Trial 28 finished with value: 0.6890889360892993 and parameters: {'bagging_fraction': 0.8933236439302734, 'bagging_freq': 3}. Best is trial 27 with value: 0.6886437067357495. bagging, val_score: 0.688644: 20%|## | 2/10 [00:03<00:13, 1.73s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001756 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688343 valid's binary_logloss: 0.689852 [200] train's binary_logloss: 0.686411 valid's binary_logloss: 0.689334 [300] train's binary_logloss: 0.685065 valid's binary_logloss: 0.688979 [400] train's binary_logloss: 0.683898 valid's binary_logloss: 0.689086 Early stopping, best iteration is: [315] train's binary_logloss: 0.684866 valid's binary_logloss: 0.688904 bagging, val_score: 0.688644: 30%|### | 3/10 [00:05<00:12, 1.73s/it][I 2020-09-27 04:56:07,585] Trial 29 finished with value: 0.6889043328094836 and parameters: {'bagging_fraction': 0.9252835167638693, 'bagging_freq': 3}. Best is trial 27 with value: 0.6886437067357495. bagging, val_score: 0.688644: 30%|### | 3/10 [00:05<00:12, 1.73s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001792 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688307 valid's binary_logloss: 0.689977 [200] train's binary_logloss: 0.686382 valid's binary_logloss: 0.689334 [300] train's binary_logloss: 0.684994 valid's binary_logloss: 0.689157 [400] train's binary_logloss: 0.683796 valid's binary_logloss: 0.688976 Early stopping, best iteration is: [399] train's binary_logloss: 0.683807 valid's binary_logloss: 0.688973 bagging, val_score: 0.688644: 40%|#### | 4/10 [00:07<00:10, 1.82s/it][I 2020-09-27 04:56:09,625] Trial 30 finished with value: 0.688972596287182 and parameters: {'bagging_fraction': 0.9070779201672944, 'bagging_freq': 3}. Best is trial 27 with value: 0.6886437067357495. bagging, val_score: 0.688644: 40%|#### | 4/10 [00:07<00:10, 1.82s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001680 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688296 valid's binary_logloss: 0.689913 [200] train's binary_logloss: 0.686352 valid's binary_logloss: 0.689474 [300] train's binary_logloss: 0.684941 valid's binary_logloss: 0.689008 [400] train's binary_logloss: 0.683731 valid's binary_logloss: 0.689001 Early stopping, best iteration is: [361] train's binary_logloss: 0.684198 valid's binary_logloss: 0.688907 bagging, val_score: 0.688644: 50%|##### | 5/10 [00:08<00:09, 1.84s/it][I 2020-09-27 04:56:11,504] Trial 31 finished with value: 0.6889071822847319 and parameters: {'bagging_fraction': 0.9123574981524005, 'bagging_freq': 3}. Best is trial 27 with value: 0.6886437067357495. bagging, val_score: 0.688644: 50%|##### | 5/10 [00:08<00:09, 1.84s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002215 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688308 valid's binary_logloss: 0.689935 [200] train's binary_logloss: 0.686414 valid's binary_logloss: 0.689366 [300] train's binary_logloss: 0.685039 valid's binary_logloss: 0.689039 [400] train's binary_logloss: 0.683854 valid's binary_logloss: 0.689061 Early stopping, best iteration is: [313] train's binary_logloss: 0.684864 valid's binary_logloss: 0.688989 bagging, val_score: 0.688644: 60%|###### | 6/10 [00:10<00:07, 1.80s/it][I 2020-09-27 04:56:13,223] Trial 32 finished with value: 0.6889888493170525 and parameters: {'bagging_fraction': 0.9247906381445511, 'bagging_freq': 3}. Best is trial 27 with value: 0.6886437067357495. bagging, val_score: 0.688644: 60%|###### | 6/10 [00:10<00:07, 1.80s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001684 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613 [200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995 [300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612 [400] train's binary_logloss: 0.683551 valid's binary_logloss: 0.688705 Early stopping, best iteration is: [330] train's binary_logloss: 0.684394 valid's binary_logloss: 0.68851 bagging, val_score: 0.688510: 70%|####### | 7/10 [00:13<00:06, 2.23s/it][I 2020-09-27 04:56:16,468] Trial 33 finished with value: 0.6885102663693616 and parameters: {'bagging_fraction': 0.7258338472094388, 'bagging_freq': 1}. Best is trial 33 with value: 0.6885102663693616. bagging, val_score: 0.688510: 70%|####### | 7/10 [00:13<00:06, 2.23s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.003707 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687986 valid's binary_logloss: 0.689762 [200] train's binary_logloss: 0.686029 valid's binary_logloss: 0.689311 [300] train's binary_logloss: 0.684604 valid's binary_logloss: 0.689125 Early stopping, best iteration is: [243] train's binary_logloss: 0.685398 valid's binary_logloss: 0.689001 bagging, val_score: 0.688510: 80%|######## | 8/10 [00:15<00:04, 2.04s/it][I 2020-09-27 04:56:18,063] Trial 34 finished with value: 0.6890013155830245 and parameters: {'bagging_fraction': 0.5748859423991363, 'bagging_freq': 1}. Best is trial 33 with value: 0.6885102663693616. bagging, val_score: 0.688510: 80%|######## | 8/10 [00:15<00:04, 2.04s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.003161 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688221 valid's binary_logloss: 0.689816 [200] train's binary_logloss: 0.686335 valid's binary_logloss: 0.688896 Early stopping, best iteration is: [196] train's binary_logloss: 0.686393 valid's binary_logloss: 0.688862 bagging, val_score: 0.688510: 90%|######### | 9/10 [00:16<00:01, 1.81s/it][I 2020-09-27 04:56:19,330] Trial 35 finished with value: 0.6888615132015019 and parameters: {'bagging_fraction': 0.7649026913055617, 'bagging_freq': 6}. Best is trial 33 with value: 0.6885102663693616. bagging, val_score: 0.688510: 90%|######### | 9/10 [00:16<00:01, 1.81s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001740 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688211 valid's binary_logloss: 0.689704 [200] train's binary_logloss: 0.686297 valid's binary_logloss: 0.689225 [300] train's binary_logloss: 0.684963 valid's binary_logloss: 0.688871 Early stopping, best iteration is: [229] train's binary_logloss: 0.685871 valid's binary_logloss: 0.688848 bagging, val_score: 0.688510: 100%|##########| 10/10 [00:18<00:00, 1.68s/it][I 2020-09-27 04:56:20,711] Trial 36 finished with value: 0.6888483568488261 and parameters: {'bagging_fraction': 0.7251434202313678, 'bagging_freq': 7}. Best is trial 33 with value: 0.6885102663693616. bagging, val_score: 0.688510: 100%|##########| 10/10 [00:18<00:00, 1.82s/it] feature_fraction_stage2, val_score: 0.688510: 0%| | 0/3 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001750 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613 [200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995 [300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612 [400] train's binary_logloss: 0.683551 valid's binary_logloss: 0.688705 Early stopping, best iteration is: [330] train's binary_logloss: 0.684394 valid's binary_logloss: 0.68851 feature_fraction_stage2, val_score: 0.688510: 33%|###3 | 1/3 [00:01<00:03, 1.59s/it][I 2020-09-27 04:56:22,315] Trial 37 finished with value: 0.6885102663693616 and parameters: {'feature_fraction': 0.9840000000000001}. Best is trial 37 with value: 0.6885102663693616. feature_fraction_stage2, val_score: 0.688510: 33%|###3 | 1/3 [00:01<00:03, 1.59s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001619 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688132 valid's binary_logloss: 0.689669 [200] train's binary_logloss: 0.686204 valid's binary_logloss: 0.689145 [300] train's binary_logloss: 0.684787 valid's binary_logloss: 0.688653 [400] train's binary_logloss: 0.683619 valid's binary_logloss: 0.688636 Early stopping, best iteration is: [392] train's binary_logloss: 0.683706 valid's binary_logloss: 0.688545 feature_fraction_stage2, val_score: 0.688510: 67%|######6 | 2/3 [00:03<00:01, 1.66s/it][I 2020-09-27 04:56:24,121] Trial 38 finished with value: 0.6885449109441385 and parameters: {'feature_fraction': 0.92}. Best is trial 37 with value: 0.6885102663693616. feature_fraction_stage2, val_score: 0.688510: 67%|######6 | 2/3 [00:03<00:01, 1.66s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001677 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68813 valid's binary_logloss: 0.689528 [200] train's binary_logloss: 0.686156 valid's binary_logloss: 0.688931 [300] train's binary_logloss: 0.684723 valid's binary_logloss: 0.688714 [400] train's binary_logloss: 0.683522 valid's binary_logloss: 0.688698 Early stopping, best iteration is: [382] train's binary_logloss: 0.683729 valid's binary_logloss: 0.688541 feature_fraction_stage2, val_score: 0.688510: 100%|##########| 3/3 [00:05<00:00, 1.69s/it][I 2020-09-27 04:56:25,892] Trial 39 finished with value: 0.688541416001024 and parameters: {'feature_fraction': 0.9520000000000001}. Best is trial 37 with value: 0.6885102663693616. feature_fraction_stage2, val_score: 0.688510: 100%|##########| 3/3 [00:05<00:00, 1.72s/it] regularization_factors, val_score: 0.688510: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.005767 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613 [200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995 [300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612 [400] train's binary_logloss: 0.683573 valid's binary_logloss: 0.68868 Early stopping, best iteration is: [380] train's binary_logloss: 0.683811 valid's binary_logloss: 0.688528 regularization_factors, val_score: 0.688510: 5%|5 | 1/20 [00:01<00:33, 1.76s/it][I 2020-09-27 04:56:27,666] Trial 40 finished with value: 0.6885276922964685 and parameters: {'lambda_l1': 1.5626678097623024e-08, 'lambda_l2': 0.0035654183554903016}. Best is trial 40 with value: 0.6885276922964685. regularization_factors, val_score: 0.688510: 5%|5 | 1/20 [00:01<00:33, 1.76s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002094 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613 [200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995 [300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612 [400] train's binary_logloss: 0.683572 valid's binary_logloss: 0.68868 Early stopping, best iteration is: [380] train's binary_logloss: 0.683811 valid's binary_logloss: 0.688528 regularization_factors, val_score: 0.688510: 10%|# | 2/20 [00:03<00:31, 1.76s/it][I 2020-09-27 04:56:29,430] Trial 41 finished with value: 0.6885276861626745 and parameters: {'lambda_l1': 3.870259482083587e-08, 'lambda_l2': 0.0031399198971774163}. Best is trial 41 with value: 0.6885276861626745. regularization_factors, val_score: 0.688510: 10%|# | 2/20 [00:03<00:31, 1.76s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001738 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613 [200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995 [300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612 [400] train's binary_logloss: 0.683572 valid's binary_logloss: 0.68868 Early stopping, best iteration is: [380] train's binary_logloss: 0.683811 valid's binary_logloss: 0.688528 regularization_factors, val_score: 0.688510: 15%|#5 | 3/20 [00:05<00:29, 1.76s/it][I 2020-09-27 04:56:31,201] Trial 42 finished with value: 0.6885276847893826 and parameters: {'lambda_l1': 1.2562118956422349e-08, 'lambda_l2': 0.0030450124632060223}. Best is trial 42 with value: 0.6885276847893826. regularization_factors, val_score: 0.688510: 15%|#5 | 3/20 [00:05<00:29, 1.76s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013796 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613 [200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995 [300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612 [400] train's binary_logloss: 0.683573 valid's binary_logloss: 0.68868 Early stopping, best iteration is: [380] train's binary_logloss: 0.683811 valid's binary_logloss: 0.688528 regularization_factors, val_score: 0.688510: 20%|## | 4/20 [00:07<00:28, 1.81s/it][I 2020-09-27 04:56:33,111] Trial 43 finished with value: 0.6885276913121847 and parameters: {'lambda_l1': 1.1600996445683937e-08, 'lambda_l2': 0.0034970196571133183}. Best is trial 42 with value: 0.6885276847893826. regularization_factors, val_score: 0.688510: 20%|## | 4/20 [00:07<00:28, 1.81s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009630 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613 [200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995 [300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612 [400] train's binary_logloss: 0.683572 valid's binary_logloss: 0.68868 Early stopping, best iteration is: [380] train's binary_logloss: 0.683811 valid's binary_logloss: 0.688528 regularization_factors, val_score: 0.688510: 25%|##5 | 5/20 [00:09<00:27, 1.84s/it][I 2020-09-27 04:56:35,011] Trial 44 finished with value: 0.6885276777308208 and parameters: {'lambda_l1': 1.5284605069192858e-08, 'lambda_l2': 0.002555456660087826}. Best is trial 44 with value: 0.6885276777308208. regularization_factors, val_score: 0.688510: 25%|##5 | 5/20 [00:09<00:27, 1.84s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001670 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613 [200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995 [300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612 [400] train's binary_logloss: 0.683572 valid's binary_logloss: 0.68868 Early stopping, best iteration is: [380] train's binary_logloss: 0.683811 valid's binary_logloss: 0.688528 regularization_factors, val_score: 0.688510: 30%|### | 6/20 [00:10<00:25, 1.81s/it][I 2020-09-27 04:56:36,753] Trial 45 finished with value: 0.6885276789659555 and parameters: {'lambda_l1': 1.761891864374289e-08, 'lambda_l2': 0.0026412614459624425}. Best is trial 44 with value: 0.6885276777308208. regularization_factors, val_score: 0.688510: 30%|### | 6/20 [00:10<00:25, 1.81s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010747 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688144 valid's binary_logloss: 0.689555 [200] train's binary_logloss: 0.686223 valid's binary_logloss: 0.688989 [300] train's binary_logloss: 0.684835 valid's binary_logloss: 0.688807 [400] train's binary_logloss: 0.683655 valid's binary_logloss: 0.688801 Early stopping, best iteration is: [366] train's binary_logloss: 0.684032 valid's binary_logloss: 0.688577 regularization_factors, val_score: 0.688510: 35%|###5 | 7/20 [00:12<00:23, 1.82s/it][I 2020-09-27 04:56:38,598] Trial 46 finished with value: 0.6885772867402516 and parameters: {'lambda_l1': 0.06434172773582825, 'lambda_l2': 4.068285530948775e-06}. Best is trial 44 with value: 0.6885276777308208. regularization_factors, val_score: 0.688510: 35%|###5 | 7/20 [00:12<00:23, 1.82s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001651 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688154 valid's binary_logloss: 0.689765 [200] train's binary_logloss: 0.686207 valid's binary_logloss: 0.689123 [300] train's binary_logloss: 0.684829 valid's binary_logloss: 0.688815 [400] train's binary_logloss: 0.683653 valid's binary_logloss: 0.688858 Early stopping, best iteration is: [378] train's binary_logloss: 0.68391 valid's binary_logloss: 0.688721 regularization_factors, val_score: 0.688510: 40%|#### | 8/20 [00:14<00:21, 1.81s/it][I 2020-09-27 04:56:40,403] Trial 47 finished with value: 0.6887213435654922 and parameters: {'lambda_l1': 1.2218484580158208e-06, 'lambda_l2': 0.5657158290313663}. Best is trial 44 with value: 0.6885276777308208. regularization_factors, val_score: 0.688510: 40%|#### | 8/20 [00:14<00:21, 1.81s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001718 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613 [200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995 [300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612 [400] train's binary_logloss: 0.683551 valid's binary_logloss: 0.688705 Early stopping, best iteration is: [330] train's binary_logloss: 0.684394 valid's binary_logloss: 0.68851 regularization_factors, val_score: 0.688510: 45%|####5 | 9/20 [00:16<00:19, 1.77s/it][I 2020-09-27 04:56:42,076] Trial 48 finished with value: 0.6885102681025668 and parameters: {'lambda_l1': 1.1710448780595842e-05, 'lambda_l2': 6.970955109724959e-05}. Best is trial 48 with value: 0.6885102681025668. regularization_factors, val_score: 0.688510: 45%|####5 | 9/20 [00:16<00:19, 1.77s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001926 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613 [200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995 [300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612 [400] train's binary_logloss: 0.683551 valid's binary_logloss: 0.688705 Early stopping, best iteration is: [330] train's binary_logloss: 0.684394 valid's binary_logloss: 0.68851 regularization_factors, val_score: 0.688510: 50%|##### | 10/20 [00:17<00:17, 1.71s/it][I 2020-09-27 04:56:43,655] Trial 49 finished with value: 0.6885102671718292 and parameters: {'lambda_l1': 2.6398830760982305e-05, 'lambda_l2': 8.347668509260632e-06}. Best is trial 49 with value: 0.6885102671718292. regularization_factors, val_score: 0.688510: 50%|##### | 10/20 [00:17<00:17, 1.71s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001747 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613 [200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995 [300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612 [400] train's binary_logloss: 0.683572 valid's binary_logloss: 0.68868 Early stopping, best iteration is: [380] train's binary_logloss: 0.683811 valid's binary_logloss: 0.688528 regularization_factors, val_score: 0.688510: 55%|#####5 | 11/20 [00:19<00:15, 1.73s/it][I 2020-09-27 04:56:45,435] Trial 50 finished with value: 0.6885276427800883 and parameters: {'lambda_l1': 9.524089932248691e-05, 'lambda_l2': 2.4636927718157655e-06}. Best is trial 49 with value: 0.6885102671718292. regularization_factors, val_score: 0.688510: 55%|#####5 | 11/20 [00:19<00:15, 1.73s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001941 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613 [200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995 [300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612 [400] train's binary_logloss: 0.683572 valid's binary_logloss: 0.68868 Early stopping, best iteration is: [380] train's binary_logloss: 0.683811 valid's binary_logloss: 0.688528 regularization_factors, val_score: 0.688510: 60%|###### | 12/20 [00:21<00:13, 1.75s/it][I 2020-09-27 04:56:47,207] Trial 51 finished with value: 0.6885276426289637 and parameters: {'lambda_l1': 8.799095559564636e-05, 'lambda_l2': 2.1427847942855786e-06}. Best is trial 49 with value: 0.6885102671718292. regularization_factors, val_score: 0.688510: 60%|###### | 12/20 [00:21<00:13, 1.75s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001695 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613 [200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995 [300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612 [400] train's binary_logloss: 0.683572 valid's binary_logloss: 0.68868 Early stopping, best iteration is: [380] train's binary_logloss: 0.683811 valid's binary_logloss: 0.688528 regularization_factors, val_score: 0.688510: 65%|######5 | 13/20 [00:23<00:12, 1.75s/it][I 2020-09-27 04:56:48,982] Trial 52 finished with value: 0.688527642704103 and parameters: {'lambda_l1': 9.263556836103949e-05, 'lambda_l2': 8.942562462981873e-07}. Best is trial 49 with value: 0.6885102671718292. regularization_factors, val_score: 0.688510: 65%|######5 | 13/20 [00:23<00:12, 1.75s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001690 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613 [200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995 [300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612 [400] train's binary_logloss: 0.683572 valid's binary_logloss: 0.68868 Early stopping, best iteration is: [380] train's binary_logloss: 0.683811 valid's binary_logloss: 0.688528 regularization_factors, val_score: 0.688510: 70%|####### | 14/20 [00:24<00:10, 1.75s/it][I 2020-09-27 04:56:50,729] Trial 53 finished with value: 0.6885276427926035 and parameters: {'lambda_l1': 9.755819246036098e-05, 'lambda_l2': 1.883877450611394e-08}. Best is trial 49 with value: 0.6885102671718292. regularization_factors, val_score: 0.688510: 70%|####### | 14/20 [00:24<00:10, 1.75s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011495 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613 [200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995 [300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612 [400] train's binary_logloss: 0.683551 valid's binary_logloss: 0.688705 Early stopping, best iteration is: [330] train's binary_logloss: 0.684394 valid's binary_logloss: 0.68851 regularization_factors, val_score: 0.688510: 75%|#######5 | 15/20 [00:26<00:08, 1.74s/it][I 2020-09-27 04:56:52,434] Trial 54 finished with value: 0.6885102665832229 and parameters: {'lambda_l1': 7.204721356400207e-06, 'lambda_l2': 1.9515753697920597e-06}. Best is trial 54 with value: 0.6885102665832229. regularization_factors, val_score: 0.688510: 75%|#######5 | 15/20 [00:26<00:08, 1.74s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001867 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613 [200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995 [300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612 [400] train's binary_logloss: 0.683551 valid's binary_logloss: 0.688705 Early stopping, best iteration is: [330] train's binary_logloss: 0.684394 valid's binary_logloss: 0.68851 regularization_factors, val_score: 0.688510: 80%|######## | 16/20 [00:28<00:06, 1.69s/it][I 2020-09-27 04:56:54,005] Trial 55 finished with value: 0.6885102672275221 and parameters: {'lambda_l1': 2.7150467223950745e-06, 'lambda_l2': 3.804591403735543e-05}. Best is trial 54 with value: 0.6885102665832229. regularization_factors, val_score: 0.688510: 80%|######## | 16/20 [00:28<00:06, 1.69s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001819 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613 [200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995 [300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612 [400] train's binary_logloss: 0.683551 valid's binary_logloss: 0.688705 Early stopping, best iteration is: [330] train's binary_logloss: 0.684394 valid's binary_logloss: 0.68851 regularization_factors, val_score: 0.688510: 85%|########5 | 17/20 [00:29<00:04, 1.67s/it][I 2020-09-27 04:56:55,619] Trial 56 finished with value: 0.6885102669786968 and parameters: {'lambda_l1': 1.4913904132624564e-06, 'lambda_l2': 2.7565544406824723e-05}. Best is trial 54 with value: 0.6885102665832229. regularization_factors, val_score: 0.688510: 85%|########5 | 17/20 [00:29<00:04, 1.67s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001735 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613 [200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995 [300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612 [400] train's binary_logloss: 0.683551 valid's binary_logloss: 0.688705 Early stopping, best iteration is: [330] train's binary_logloss: 0.684394 valid's binary_logloss: 0.68851 regularization_factors, val_score: 0.688510: 90%|######### | 18/20 [00:31<00:03, 1.63s/it][I 2020-09-27 04:56:57,171] Trial 57 finished with value: 0.6885102674416176 and parameters: {'lambda_l1': 7.533020720507337e-07, 'lambda_l2': 5.04819339961594e-05}. Best is trial 54 with value: 0.6885102665832229. regularization_factors, val_score: 0.688510: 90%|######### | 18/20 [00:31<00:03, 1.63s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001739 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613 [200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995 [300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688612 [400] train's binary_logloss: 0.683551 valid's binary_logloss: 0.688705 Early stopping, best iteration is: [330] train's binary_logloss: 0.684394 valid's binary_logloss: 0.68851 regularization_factors, val_score: 0.688510: 95%|#########5| 19/20 [00:32<00:01, 1.63s/it][I 2020-09-27 04:56:58,787] Trial 58 finished with value: 0.6885102664101546 and parameters: {'lambda_l1': 1.6489132998257337e-06, 'lambda_l2': 7.130405244958287e-08}. Best is trial 58 with value: 0.6885102664101546. regularization_factors, val_score: 0.688510: 95%|#########5| 19/20 [00:32<00:01, 1.63s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001734 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613 [200] train's binary_logloss: 0.68616 valid's binary_logloss: 0.688995 [300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688613 [400] train's binary_logloss: 0.683573 valid's binary_logloss: 0.68868 Early stopping, best iteration is: [380] train's binary_logloss: 0.683811 valid's binary_logloss: 0.688528 regularization_factors, val_score: 0.688510: 100%|##########| 20/20 [00:34<00:00, 1.67s/it][I 2020-09-27 04:57:00,567] Trial 59 finished with value: 0.6885278039688245 and parameters: {'lambda_l1': 0.008341325209129346, 'lambda_l2': 6.068578447774457e-08}. Best is trial 58 with value: 0.6885102664101546. regularization_factors, val_score: 0.688510: 100%|##########| 20/20 [00:34<00:00, 1.73s/it] min_data_in_leaf, val_score: 0.688510: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002224 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613 [200] train's binary_logloss: 0.686175 valid's binary_logloss: 0.68904 [300] train's binary_logloss: 0.684779 valid's binary_logloss: 0.68875 [400] train's binary_logloss: 0.683563 valid's binary_logloss: 0.68872 [500] train's binary_logloss: 0.682433 valid's binary_logloss: 0.688523 [600] train's binary_logloss: 0.681362 valid's binary_logloss: 0.688598 Early stopping, best iteration is: [540] train's binary_logloss: 0.682021 valid's binary_logloss: 0.688417 min_data_in_leaf, val_score: 0.688417: 20%|## | 1/5 [00:02<00:09, 2.34s/it][I 2020-09-27 04:57:02,917] Trial 60 finished with value: 0.6884174980111903 and parameters: {'min_child_samples': 25}. Best is trial 60 with value: 0.6884174980111903. min_data_in_leaf, val_score: 0.688417: 20%|## | 1/5 [00:02<00:09, 2.34s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001678 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613 [200] train's binary_logloss: 0.686186 valid's binary_logloss: 0.68902 [300] train's binary_logloss: 0.684775 valid's binary_logloss: 0.688899 [400] train's binary_logloss: 0.683559 valid's binary_logloss: 0.688863 Early stopping, best iteration is: [378] train's binary_logloss: 0.68381 valid's binary_logloss: 0.68872 min_data_in_leaf, val_score: 0.688417: 40%|#### | 2/5 [00:04<00:06, 2.16s/it][I 2020-09-27 04:57:04,665] Trial 61 finished with value: 0.6887198054735056 and parameters: {'min_child_samples': 50}. Best is trial 60 with value: 0.6884174980111903. min_data_in_leaf, val_score: 0.688417: 40%|#### | 2/5 [00:04<00:06, 2.16s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001665 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688189 valid's binary_logloss: 0.689923 [200] train's binary_logloss: 0.686316 valid's binary_logloss: 0.689183 [300] train's binary_logloss: 0.684943 valid's binary_logloss: 0.689041 Early stopping, best iteration is: [254] train's binary_logloss: 0.685519 valid's binary_logloss: 0.688957 min_data_in_leaf, val_score: 0.688417: 60%|###### | 3/5 [00:05<00:03, 1.91s/it][I 2020-09-27 04:57:05,984] Trial 62 finished with value: 0.6889570125436887 and parameters: {'min_child_samples': 100}. Best is trial 60 with value: 0.6884174980111903. min_data_in_leaf, val_score: 0.688417: 60%|###### | 3/5 [00:05<00:03, 1.91s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001746 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613 [200] train's binary_logloss: 0.686178 valid's binary_logloss: 0.688982 [300] train's binary_logloss: 0.684774 valid's binary_logloss: 0.688717 [400] train's binary_logloss: 0.683552 valid's binary_logloss: 0.688696 Early stopping, best iteration is: [383] train's binary_logloss: 0.683767 valid's binary_logloss: 0.688615 min_data_in_leaf, val_score: 0.688417: 80%|######## | 4/5 [00:07<00:01, 1.86s/it][I 2020-09-27 04:57:07,748] Trial 63 finished with value: 0.6886149362033982 and parameters: {'min_child_samples': 10}. Best is trial 60 with value: 0.6884174980111903. min_data_in_leaf, val_score: 0.688417: 80%|######## | 4/5 [00:07<00:01, 1.86s/it][LightGBM] [Info] Number of positive: 46664, number of negative: 46362 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001731 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4688 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.501623 -> initscore=0.006493 [LightGBM] [Info] Start training from score 0.006493 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688124 valid's binary_logloss: 0.689613 [200] train's binary_logloss: 0.686183 valid's binary_logloss: 0.688973 [300] train's binary_logloss: 0.684776 valid's binary_logloss: 0.688751 [400] train's binary_logloss: 0.683531 valid's binary_logloss: 0.688863 Early stopping, best iteration is: [306] train's binary_logloss: 0.684696 valid's binary_logloss: 0.688722 min_data_in_leaf, val_score: 0.688417: 100%|##########| 5/5 [00:08<00:00, 1.76s/it][I 2020-09-27 04:57:09,249] Trial 64 finished with value: 0.6887218892158072 and parameters: {'min_child_samples': 5}. Best is trial 60 with value: 0.6884174980111903. min_data_in_leaf, val_score: 0.688417: 100%|##########| 5/5 [00:08<00:00, 1.73s/it] Fold : 5 [I 2020-09-27 04:57:09,427] A new study created in memory with name: no-name-1b7c624a-f0b2-4aea-8fc8-aa6993df7768 feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001589 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.662759 valid's binary_logloss: 0.691799 Early stopping, best iteration is: [36] train's binary_logloss: 0.678598 valid's binary_logloss: 0.69056 feature_fraction, val_score: 0.690560: 14%|#4 | 1/7 [00:00<00:05, 1.06it/s][I 2020-09-27 04:57:10,381] Trial 0 finished with value: 0.6905601497910684 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 0 with value: 0.6905601497910684. feature_fraction, val_score: 0.690560: 14%|#4 | 1/7 [00:00<00:05, 1.06it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.007342 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.66369 valid's binary_logloss: 0.691565 Early stopping, best iteration is: [25] train's binary_logloss: 0.68226 valid's binary_logloss: 0.690559 feature_fraction, val_score: 0.690559: 29%|##8 | 2/7 [00:01<00:04, 1.15it/s][I 2020-09-27 04:57:11,086] Trial 1 finished with value: 0.6905590100900298 and parameters: {'feature_fraction': 0.7}. Best is trial 1 with value: 0.6905590100900298. feature_fraction, val_score: 0.690559: 29%|##8 | 2/7 [00:01<00:04, 1.15it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011035 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664057 valid's binary_logloss: 0.690869 Early stopping, best iteration is: [40] train's binary_logloss: 0.678145 valid's binary_logloss: 0.690182 feature_fraction, val_score: 0.690182: 43%|####2 | 3/7 [00:02<00:03, 1.23it/s][I 2020-09-27 04:57:11,765] Trial 2 finished with value: 0.6901820889991936 and parameters: {'feature_fraction': 0.6}. Best is trial 2 with value: 0.6901820889991936. feature_fraction, val_score: 0.690182: 43%|####2 | 3/7 [00:02<00:03, 1.23it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001692 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664852 valid's binary_logloss: 0.691416 Early stopping, best iteration is: [35] train's binary_logloss: 0.679773 valid's binary_logloss: 0.690316 feature_fraction, val_score: 0.690182: 57%|#####7 | 4/7 [00:03<00:02, 1.29it/s][I 2020-09-27 04:57:12,459] Trial 3 finished with value: 0.690316333441943 and parameters: {'feature_fraction': 0.5}. Best is trial 2 with value: 0.6901820889991936. feature_fraction, val_score: 0.690182: 57%|#####7 | 4/7 [00:03<00:02, 1.29it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.006609 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663021 valid's binary_logloss: 0.690361 Early stopping, best iteration is: [48] train's binary_logloss: 0.675428 valid's binary_logloss: 0.689984 feature_fraction, val_score: 0.689984: 71%|#######1 | 5/7 [00:03<00:01, 1.27it/s][I 2020-09-27 04:57:13,274] Trial 4 finished with value: 0.6899844980770151 and parameters: {'feature_fraction': 0.8}. Best is trial 4 with value: 0.6899844980770151. feature_fraction, val_score: 0.689984: 71%|#######1 | 5/7 [00:03<00:01, 1.27it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001693 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.662006 valid's binary_logloss: 0.691157 Early stopping, best iteration is: [40] train's binary_logloss: 0.67696 valid's binary_logloss: 0.690385 feature_fraction, val_score: 0.689984: 86%|########5 | 6/7 [00:04<00:00, 1.26it/s][I 2020-09-27 04:57:14,069] Trial 5 finished with value: 0.690384793694187 and parameters: {'feature_fraction': 1.0}. Best is trial 4 with value: 0.6899844980770151. feature_fraction, val_score: 0.689984: 86%|########5 | 6/7 [00:04<00:00, 1.26it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000701 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.666045 valid's binary_logloss: 0.691233 Early stopping, best iteration is: [38] train's binary_logloss: 0.67965 valid's binary_logloss: 0.690765 feature_fraction, val_score: 0.689984: 100%|##########| 7/7 [00:05<00:00, 1.31it/s][I 2020-09-27 04:57:14,764] Trial 6 finished with value: 0.6907650282143057 and parameters: {'feature_fraction': 0.4}. Best is trial 4 with value: 0.6899844980770151. feature_fraction, val_score: 0.689984: 100%|##########| 7/7 [00:05<00:00, 1.31it/s] num_leaves, val_score: 0.689984: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008471 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.532446 valid's binary_logloss: 0.701071 Early stopping, best iteration is: [12] train's binary_logloss: 0.663644 valid's binary_logloss: 0.692101 num_leaves, val_score: 0.689984: 5%|5 | 1/20 [00:01<00:24, 1.28s/it][I 2020-09-27 04:57:16,058] Trial 7 finished with value: 0.6921010909265272 and parameters: {'num_leaves': 246}. Best is trial 7 with value: 0.6921010909265272. num_leaves, val_score: 0.689984: 5%|5 | 1/20 [00:01<00:24, 1.28s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011999 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.541655 valid's binary_logloss: 0.698738 Early stopping, best iteration is: [19] train's binary_logloss: 0.651466 valid's binary_logloss: 0.691604 num_leaves, val_score: 0.689984: 10%|# | 2/20 [00:02<00:23, 1.29s/it][I 2020-09-27 04:57:17,368] Trial 8 finished with value: 0.6916040591247711 and parameters: {'num_leaves': 228}. Best is trial 8 with value: 0.6916040591247711. num_leaves, val_score: 0.689984: 10%|# | 2/20 [00:02<00:23, 1.29s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001462 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.528502 valid's binary_logloss: 0.699513 Early stopping, best iteration is: [12] train's binary_logloss: 0.662669 valid's binary_logloss: 0.69248 num_leaves, val_score: 0.689984: 15%|#5 | 3/20 [00:04<00:22, 1.34s/it][I 2020-09-27 04:57:18,834] Trial 9 finished with value: 0.6924803382332924 and parameters: {'num_leaves': 254}. Best is trial 8 with value: 0.6916040591247711. num_leaves, val_score: 0.689984: 15%|#5 | 3/20 [00:04<00:22, 1.34s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001476 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.683804 valid's binary_logloss: 0.690293 [200] train's binary_logloss: 0.67902 valid's binary_logloss: 0.690554 Early stopping, best iteration is: [108] train's binary_logloss: 0.683362 valid's binary_logloss: 0.690199 num_leaves, val_score: 0.689984: 20%|## | 4/20 [00:04<00:19, 1.19s/it][I 2020-09-27 04:57:19,664] Trial 10 finished with value: 0.6901988057794263 and parameters: {'num_leaves': 7}. Best is trial 10 with value: 0.6901988057794263. num_leaves, val_score: 0.689984: 20%|## | 4/20 [00:04<00:19, 1.19s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001573 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.677338 valid's binary_logloss: 0.690393 [200] train's binary_logloss: 0.667531 valid's binary_logloss: 0.69092 Early stopping, best iteration is: [138] train's binary_logloss: 0.673519 valid's binary_logloss: 0.690353 num_leaves, val_score: 0.689984: 25%|##5 | 5/20 [00:05<00:17, 1.15s/it][I 2020-09-27 04:57:20,726] Trial 11 finished with value: 0.6903529337197906 and parameters: {'num_leaves': 14}. Best is trial 10 with value: 0.6901988057794263. num_leaves, val_score: 0.689984: 25%|##5 | 5/20 [00:05<00:17, 1.15s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001618 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.607789 valid's binary_logloss: 0.695 Early stopping, best iteration is: [16] train's binary_logloss: 0.673018 valid's binary_logloss: 0.691131 num_leaves, val_score: 0.689984: 30%|### | 6/20 [00:06<00:15, 1.12s/it][I 2020-09-27 04:57:21,759] Trial 12 finished with value: 0.6911309288725374 and parameters: {'num_leaves': 110}. Best is trial 10 with value: 0.6901988057794263. num_leaves, val_score: 0.689984: 30%|### | 6/20 [00:06<00:15, 1.12s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002136 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.584924 valid's binary_logloss: 0.695787 Early stopping, best iteration is: [13] train's binary_logloss: 0.672012 valid's binary_logloss: 0.691214 num_leaves, val_score: 0.689984: 35%|###5 | 7/20 [00:08<00:14, 1.12s/it][I 2020-09-27 04:57:22,903] Trial 13 finished with value: 0.6912141405715722 and parameters: {'num_leaves': 148}. Best is trial 10 with value: 0.6901988057794263. num_leaves, val_score: 0.689984: 35%|###5 | 7/20 [00:08<00:14, 1.12s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001517 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.632471 valid's binary_logloss: 0.693122 Early stopping, best iteration is: [14] train's binary_logloss: 0.680218 valid's binary_logloss: 0.691475 num_leaves, val_score: 0.689984: 40%|#### | 8/20 [00:08<00:12, 1.03s/it][I 2020-09-27 04:57:23,723] Trial 14 finished with value: 0.6914751164482753 and parameters: {'num_leaves': 72}. Best is trial 10 with value: 0.6901988057794263. num_leaves, val_score: 0.689984: 40%|#### | 8/20 [00:08<00:12, 1.03s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001436 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.563691 valid's binary_logloss: 0.697937 Early stopping, best iteration is: [9] train's binary_logloss: 0.67486 valid's binary_logloss: 0.691909 num_leaves, val_score: 0.689984: 45%|####5 | 9/20 [00:10<00:11, 1.09s/it][I 2020-09-27 04:57:24,931] Trial 15 finished with value: 0.6919085281880217 and parameters: {'num_leaves': 185}. Best is trial 10 with value: 0.6901988057794263. num_leaves, val_score: 0.689984: 45%|####5 | 9/20 [00:10<00:11, 1.09s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001412 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.658128 valid's binary_logloss: 0.690955 Early stopping, best iteration is: [30] train's binary_logloss: 0.678602 valid's binary_logloss: 0.690303 num_leaves, val_score: 0.689984: 50%|##### | 10/20 [00:10<00:09, 1.02it/s][I 2020-09-27 04:57:25,662] Trial 16 finished with value: 0.6903025101431474 and parameters: {'num_leaves': 37}. Best is trial 10 with value: 0.6901988057794263. num_leaves, val_score: 0.689984: 50%|##### | 10/20 [00:10<00:09, 1.02it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011796 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.562179 valid's binary_logloss: 0.697416 Early stopping, best iteration is: [13] train's binary_logloss: 0.667864 valid's binary_logloss: 0.691752 num_leaves, val_score: 0.689984: 55%|#####5 | 11/20 [00:12<00:09, 1.04s/it][I 2020-09-27 04:57:26,832] Trial 17 finished with value: 0.6917519955746995 and parameters: {'num_leaves': 187}. Best is trial 10 with value: 0.6901988057794263. num_leaves, val_score: 0.689984: 55%|#####5 | 11/20 [00:12<00:09, 1.04s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002032 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.614951 valid's binary_logloss: 0.693854 Early stopping, best iteration is: [26] train's binary_logloss: 0.665369 valid's binary_logloss: 0.690866 num_leaves, val_score: 0.689984: 60%|###### | 12/20 [00:13<00:08, 1.04s/it][I 2020-09-27 04:57:27,890] Trial 18 finished with value: 0.6908659763700707 and parameters: {'num_leaves': 99}. Best is trial 10 with value: 0.6901988057794263. num_leaves, val_score: 0.689984: 60%|###### | 12/20 [00:13<00:08, 1.04s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001505 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.637367 valid's binary_logloss: 0.693163 Early stopping, best iteration is: [32] train's binary_logloss: 0.669343 valid's binary_logloss: 0.690841 num_leaves, val_score: 0.689984: 65%|######5 | 13/20 [00:14<00:06, 1.00it/s][I 2020-09-27 04:57:28,782] Trial 19 finished with value: 0.6908408834731546 and parameters: {'num_leaves': 66}. Best is trial 10 with value: 0.6901988057794263. num_leaves, val_score: 0.689984: 65%|######5 | 13/20 [00:14<00:06, 1.00it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008072 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.575818 valid's binary_logloss: 0.696901 Early stopping, best iteration is: [13] train's binary_logloss: 0.669822 valid's binary_logloss: 0.691657 num_leaves, val_score: 0.689984: 70%|####### | 14/20 [00:15<00:06, 1.02s/it][I 2020-09-27 04:57:29,853] Trial 20 finished with value: 0.6916567124710575 and parameters: {'num_leaves': 166}. Best is trial 10 with value: 0.6901988057794263. num_leaves, val_score: 0.689984: 70%|####### | 14/20 [00:15<00:06, 1.02s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008152 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687062 valid's binary_logloss: 0.690297 [200] train's binary_logloss: 0.684539 valid's binary_logloss: 0.690125 Early stopping, best iteration is: [195] train's binary_logloss: 0.684633 valid's binary_logloss: 0.690069 num_leaves, val_score: 0.689984: 75%|#######5 | 15/20 [00:16<00:05, 1.04s/it][I 2020-09-27 04:57:30,950] Trial 21 finished with value: 0.6900688079787631 and parameters: {'num_leaves': 4}. Best is trial 21 with value: 0.6900688079787631. num_leaves, val_score: 0.689984: 75%|#######5 | 15/20 [00:16<00:05, 1.04s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002005 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.660627 valid's binary_logloss: 0.690947 Early stopping, best iteration is: [34] train's binary_logloss: 0.678295 valid's binary_logloss: 0.690073 num_leaves, val_score: 0.689984: 80%|######## | 16/20 [00:16<00:03, 1.04it/s][I 2020-09-27 04:57:31,722] Trial 22 finished with value: 0.6900727900021961 and parameters: {'num_leaves': 34}. Best is trial 21 with value: 0.6900688079787631. num_leaves, val_score: 0.689984: 80%|######## | 16/20 [00:16<00:03, 1.04it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001577 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.653581 valid's binary_logloss: 0.692192 Early stopping, best iteration is: [26] train's binary_logloss: 0.67856 valid's binary_logloss: 0.690524 num_leaves, val_score: 0.689984: 85%|########5 | 17/20 [00:17<00:02, 1.10it/s][I 2020-09-27 04:57:32,506] Trial 23 finished with value: 0.6905242489383265 and parameters: {'num_leaves': 43}. Best is trial 21 with value: 0.6900688079787631. num_leaves, val_score: 0.689984: 85%|########5 | 17/20 [00:17<00:02, 1.10it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001600 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.682861 valid's binary_logloss: 0.690265 Early stopping, best iteration is: [76] train's binary_logloss: 0.684408 valid's binary_logloss: 0.690209 num_leaves, val_score: 0.689984: 90%|######### | 18/20 [00:18<00:01, 1.16it/s][I 2020-09-27 04:57:33,262] Trial 24 finished with value: 0.6902091508032887 and parameters: {'num_leaves': 8}. Best is trial 21 with value: 0.6900688079787631. num_leaves, val_score: 0.689984: 90%|######### | 18/20 [00:18<00:01, 1.16it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.014414 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.656074 valid's binary_logloss: 0.691055 Early stopping, best iteration is: [34] train's binary_logloss: 0.676266 valid's binary_logloss: 0.690265 num_leaves, val_score: 0.689984: 95%|#########5| 19/20 [00:19<00:00, 1.19it/s][I 2020-09-27 04:57:34,040] Trial 25 finished with value: 0.6902649750375632 and parameters: {'num_leaves': 40}. Best is trial 21 with value: 0.6900688079787631. num_leaves, val_score: 0.689984: 95%|#########5| 19/20 [00:19<00:00, 1.19it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001563 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.630456 valid's binary_logloss: 0.693 Early stopping, best iteration is: [21] train's binary_logloss: 0.67434 valid's binary_logloss: 0.691243 num_leaves, val_score: 0.689984: 100%|##########| 20/20 [00:20<00:00, 1.17it/s][I 2020-09-27 04:57:34,925] Trial 26 finished with value: 0.6912431590178266 and parameters: {'num_leaves': 75}. Best is trial 21 with value: 0.6900688079787631. num_leaves, val_score: 0.689984: 100%|##########| 20/20 [00:20<00:00, 1.01s/it] bagging, val_score: 0.689984: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001562 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.662713 valid's binary_logloss: 0.690614 Early stopping, best iteration is: [61] train's binary_logloss: 0.672055 valid's binary_logloss: 0.690011 bagging, val_score: 0.689984: 10%|# | 1/10 [00:01<00:09, 1.06s/it][I 2020-09-27 04:57:35,998] Trial 27 finished with value: 0.6900113794489978 and parameters: {'bagging_fraction': 0.7744675089831714, 'bagging_freq': 3}. Best is trial 27 with value: 0.6900113794489978. bagging, val_score: 0.689984: 10%|# | 1/10 [00:01<00:09, 1.06s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001550 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663033 valid's binary_logloss: 0.691575 Early stopping, best iteration is: [65] train's binary_logloss: 0.671129 valid's binary_logloss: 0.690781 bagging, val_score: 0.689984: 20%|## | 2/10 [00:02<00:08, 1.10s/it][I 2020-09-27 04:57:37,192] Trial 28 finished with value: 0.6907811511387683 and parameters: {'bagging_fraction': 0.7835892578642625, 'bagging_freq': 3}. Best is trial 27 with value: 0.6900113794489978. bagging, val_score: 0.689984: 20%|## | 2/10 [00:02<00:08, 1.10s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009896 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664095 valid's binary_logloss: 0.692608 Early stopping, best iteration is: [35] train's binary_logloss: 0.67955 valid's binary_logloss: 0.690369 bagging, val_score: 0.689984: 30%|### | 3/10 [00:03<00:07, 1.05s/it][I 2020-09-27 04:57:38,119] Trial 29 finished with value: 0.6903685689943968 and parameters: {'bagging_fraction': 0.4766066067413073, 'bagging_freq': 1}. Best is trial 27 with value: 0.6900113794489978. bagging, val_score: 0.689984: 30%|### | 3/10 [00:03<00:07, 1.05s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008136 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.662586 valid's binary_logloss: 0.691009 Early stopping, best iteration is: [26] train's binary_logloss: 0.68167 valid's binary_logloss: 0.690305 bagging, val_score: 0.689984: 40%|#### | 4/10 [00:03<00:05, 1.02it/s][I 2020-09-27 04:57:38,929] Trial 30 finished with value: 0.6903053053602763 and parameters: {'bagging_fraction': 0.9933366847840979, 'bagging_freq': 7}. Best is trial 27 with value: 0.6900113794489978. bagging, val_score: 0.689984: 40%|#### | 4/10 [00:03<00:05, 1.02it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001593 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663295 valid's binary_logloss: 0.692575 Early stopping, best iteration is: [46] train's binary_logloss: 0.676198 valid's binary_logloss: 0.690885 bagging, val_score: 0.689984: 50%|##### | 5/10 [00:04<00:04, 1.04it/s][I 2020-09-27 04:57:39,840] Trial 31 finished with value: 0.6908846966555133 and parameters: {'bagging_fraction': 0.7357347346451741, 'bagging_freq': 4}. Best is trial 27 with value: 0.6900113794489978. bagging, val_score: 0.689984: 50%|##### | 5/10 [00:04<00:04, 1.04it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001590 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.662926 valid's binary_logloss: 0.691444 Early stopping, best iteration is: [22] train's binary_logloss: 0.683062 valid's binary_logloss: 0.690387 bagging, val_score: 0.689984: 60%|###### | 6/10 [00:05<00:03, 1.09it/s][I 2020-09-27 04:57:40,659] Trial 32 finished with value: 0.6903865124235224 and parameters: {'bagging_fraction': 0.9606439169637138, 'bagging_freq': 4}. Best is trial 27 with value: 0.6900113794489978. bagging, val_score: 0.689984: 60%|###### | 6/10 [00:05<00:03, 1.09it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001535 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.66338 valid's binary_logloss: 0.692174 Early stopping, best iteration is: [35] train's binary_logloss: 0.679143 valid's binary_logloss: 0.690222 bagging, val_score: 0.689984: 70%|####### | 7/10 [00:06<00:02, 1.15it/s][I 2020-09-27 04:57:41,428] Trial 33 finished with value: 0.6902221405606176 and parameters: {'bagging_fraction': 0.5501404306458005, 'bagging_freq': 1}. Best is trial 27 with value: 0.6900113794489978. bagging, val_score: 0.689984: 70%|####### | 7/10 [00:06<00:02, 1.15it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001493 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.662525 valid's binary_logloss: 0.691554 Early stopping, best iteration is: [30] train's binary_logloss: 0.680411 valid's binary_logloss: 0.690703 bagging, val_score: 0.689984: 80%|######## | 8/10 [00:07<00:01, 1.16it/s][I 2020-09-27 04:57:42,266] Trial 34 finished with value: 0.6907033298544214 and parameters: {'bagging_fraction': 0.8542206886051794, 'bagging_freq': 6}. Best is trial 27 with value: 0.6900113794489978. bagging, val_score: 0.689984: 80%|######## | 8/10 [00:07<00:01, 1.16it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011907 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663364 valid's binary_logloss: 0.69162 Early stopping, best iteration is: [35] train's binary_logloss: 0.679523 valid's binary_logloss: 0.690026 bagging, val_score: 0.689984: 90%|######### | 9/10 [00:08<00:00, 1.18it/s][I 2020-09-27 04:57:43,075] Trial 35 finished with value: 0.6900257534785464 and parameters: {'bagging_fraction': 0.6072769654268587, 'bagging_freq': 3}. Best is trial 27 with value: 0.6900113794489978. bagging, val_score: 0.689984: 90%|######### | 9/10 [00:08<00:00, 1.18it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001565 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663332 valid's binary_logloss: 0.691011 Early stopping, best iteration is: [30] train's binary_logloss: 0.680985 valid's binary_logloss: 0.69027 bagging, val_score: 0.689984: 100%|##########| 10/10 [00:08<00:00, 1.19it/s][I 2020-09-27 04:57:43,910] Trial 36 finished with value: 0.6902695334181121 and parameters: {'bagging_fraction': 0.618007946081386, 'bagging_freq': 3}. Best is trial 27 with value: 0.6900113794489978. bagging, val_score: 0.689984: 100%|##########| 10/10 [00:08<00:00, 1.11it/s] feature_fraction_stage2, val_score: 0.689984: 0%| | 0/6 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013405 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.662852 valid's binary_logloss: 0.691261 Early stopping, best iteration is: [51] train's binary_logloss: 0.674585 valid's binary_logloss: 0.690386 feature_fraction_stage2, val_score: 0.689984: 17%|#6 | 1/6 [00:00<00:04, 1.23it/s][I 2020-09-27 04:57:44,733] Trial 37 finished with value: 0.690385946557375 and parameters: {'feature_fraction': 0.7520000000000001}. Best is trial 37 with value: 0.690385946557375. feature_fraction_stage2, val_score: 0.689984: 17%|#6 | 1/6 [00:00<00:04, 1.23it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008459 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663021 valid's binary_logloss: 0.690361 Early stopping, best iteration is: [48] train's binary_logloss: 0.675428 valid's binary_logloss: 0.689984 feature_fraction_stage2, val_score: 0.689984: 33%|###3 | 2/6 [00:01<00:03, 1.23it/s][I 2020-09-27 04:57:45,549] Trial 38 finished with value: 0.6899844980770151 and parameters: {'feature_fraction': 0.8160000000000001}. Best is trial 38 with value: 0.6899844980770151. feature_fraction_stage2, val_score: 0.689984: 33%|###3 | 2/6 [00:01<00:03, 1.23it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001668 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.662759 valid's binary_logloss: 0.691799 Early stopping, best iteration is: [36] train's binary_logloss: 0.678598 valid's binary_logloss: 0.69056 feature_fraction_stage2, val_score: 0.689984: 50%|##### | 3/6 [00:02<00:02, 1.26it/s][I 2020-09-27 04:57:46,293] Trial 39 finished with value: 0.6905601497910684 and parameters: {'feature_fraction': 0.88}. Best is trial 38 with value: 0.6899844980770151. feature_fraction_stage2, val_score: 0.689984: 50%|##### | 3/6 [00:02<00:02, 1.26it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008728 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.662852 valid's binary_logloss: 0.691261 Early stopping, best iteration is: [51] train's binary_logloss: 0.674585 valid's binary_logloss: 0.690386 feature_fraction_stage2, val_score: 0.689984: 67%|######6 | 4/6 [00:03<00:01, 1.27it/s][I 2020-09-27 04:57:47,065] Trial 40 finished with value: 0.6903859465573751 and parameters: {'feature_fraction': 0.784}. Best is trial 38 with value: 0.6899844980770151. feature_fraction_stage2, val_score: 0.689984: 67%|######6 | 4/6 [00:03<00:01, 1.27it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013482 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663303 valid's binary_logloss: 0.690672 Early stopping, best iteration is: [53] train's binary_logloss: 0.674298 valid's binary_logloss: 0.690024 feature_fraction_stage2, val_score: 0.689984: 83%|########3 | 5/6 [00:03<00:00, 1.27it/s][I 2020-09-27 04:57:47,852] Trial 41 finished with value: 0.6900236664988709 and parameters: {'feature_fraction': 0.7200000000000001}. Best is trial 38 with value: 0.6899844980770151. feature_fraction_stage2, val_score: 0.689984: 83%|########3 | 5/6 [00:03<00:00, 1.27it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009074 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.662643 valid's binary_logloss: 0.690983 Early stopping, best iteration is: [32] train's binary_logloss: 0.679699 valid's binary_logloss: 0.690557 feature_fraction_stage2, val_score: 0.689984: 100%|##########| 6/6 [00:04<00:00, 1.31it/s][I 2020-09-27 04:57:48,572] Trial 42 finished with value: 0.6905568387008639 and parameters: {'feature_fraction': 0.8480000000000001}. Best is trial 38 with value: 0.6899844980770151. feature_fraction_stage2, val_score: 0.689984: 100%|##########| 6/6 [00:04<00:00, 1.29it/s] regularization_factors, val_score: 0.689984: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001489 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663071 valid's binary_logloss: 0.690424 Early stopping, best iteration is: [79] train's binary_logloss: 0.667888 valid's binary_logloss: 0.689968 regularization_factors, val_score: 0.689968: 5%|5 | 1/20 [00:01<00:19, 1.00s/it][I 2020-09-27 04:57:49,589] Trial 43 finished with value: 0.6899675224070905 and parameters: {'lambda_l1': 0.0034339426266632865, 'lambda_l2': 2.9528148173408035e-06}. Best is trial 43 with value: 0.6899675224070905. regularization_factors, val_score: 0.689968: 5%|5 | 1/20 [00:01<00:19, 1.00s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001532 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.66306 valid's binary_logloss: 0.690447 Early stopping, best iteration is: [79] train's binary_logloss: 0.66789 valid's binary_logloss: 0.689967 regularization_factors, val_score: 0.689967: 10%|# | 2/20 [00:02<00:18, 1.01s/it][I 2020-09-27 04:57:50,613] Trial 44 finished with value: 0.6899674411707333 and parameters: {'lambda_l1': 0.004662561477422725, 'lambda_l2': 1.3948625139125175e-06}. Best is trial 44 with value: 0.6899674411707333. regularization_factors, val_score: 0.689967: 10%|# | 2/20 [00:02<00:18, 1.01s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.003785 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663061 valid's binary_logloss: 0.690447 Early stopping, best iteration is: [79] train's binary_logloss: 0.66789 valid's binary_logloss: 0.689967 regularization_factors, val_score: 0.689967: 15%|#5 | 3/20 [00:03<00:17, 1.01s/it][I 2020-09-27 04:57:51,643] Trial 45 finished with value: 0.6899674025098815 and parameters: {'lambda_l1': 0.00524640370808824, 'lambda_l2': 1.4398621808620703e-06}. Best is trial 45 with value: 0.6899674025098815. regularization_factors, val_score: 0.689967: 15%|#5 | 3/20 [00:03<00:17, 1.01s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008016 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.662786 valid's binary_logloss: 0.690799 Early stopping, best iteration is: [47] train's binary_logloss: 0.67568 valid's binary_logloss: 0.69003 regularization_factors, val_score: 0.689967: 20%|## | 4/20 [00:03<00:15, 1.05it/s][I 2020-09-27 04:57:52,442] Trial 46 finished with value: 0.6900296520521778 and parameters: {'lambda_l1': 0.010031342307687478, 'lambda_l2': 8.690635159166719e-07}. Best is trial 45 with value: 0.6899674025098815. regularization_factors, val_score: 0.689967: 20%|## | 4/20 [00:03<00:15, 1.05it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001557 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.66307 valid's binary_logloss: 0.690424 Early stopping, best iteration is: [79] train's binary_logloss: 0.667887 valid's binary_logloss: 0.689968 regularization_factors, val_score: 0.689967: 25%|##5 | 5/20 [00:04<00:14, 1.05it/s][I 2020-09-27 04:57:53,400] Trial 47 finished with value: 0.6899675786092242 and parameters: {'lambda_l1': 0.002586041944855055, 'lambda_l2': 2.63283121364215e-06}. Best is trial 45 with value: 0.6899674025098815. regularization_factors, val_score: 0.689967: 25%|##5 | 5/20 [00:04<00:14, 1.05it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008047 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663059 valid's binary_logloss: 0.690447 Early stopping, best iteration is: [79] train's binary_logloss: 0.667889 valid's binary_logloss: 0.689967 regularization_factors, val_score: 0.689967: 30%|### | 6/20 [00:05<00:13, 1.05it/s][I 2020-09-27 04:57:54,339] Trial 48 finished with value: 0.6899674839198124 and parameters: {'lambda_l1': 0.004015863099526828, 'lambda_l2': 2.280356884231991e-06}. Best is trial 45 with value: 0.6899674025098815. regularization_factors, val_score: 0.689967: 30%|### | 6/20 [00:05<00:13, 1.05it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001644 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663071 valid's binary_logloss: 0.690424 Early stopping, best iteration is: [79] train's binary_logloss: 0.667888 valid's binary_logloss: 0.689968 regularization_factors, val_score: 0.689967: 35%|###5 | 7/20 [00:06<00:12, 1.01it/s][I 2020-09-27 04:57:55,418] Trial 49 finished with value: 0.6899675288047956 and parameters: {'lambda_l1': 0.0033375593376080356, 'lambda_l2': 2.830642312376846e-06}. Best is trial 45 with value: 0.6899674025098815. regularization_factors, val_score: 0.689967: 35%|###5 | 7/20 [00:06<00:12, 1.01it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001601 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663071 valid's binary_logloss: 0.690424 Early stopping, best iteration is: [79] train's binary_logloss: 0.667888 valid's binary_logloss: 0.689968 regularization_factors, val_score: 0.689967: 40%|#### | 8/20 [00:07<00:12, 1.00s/it][I 2020-09-27 04:57:56,448] Trial 50 finished with value: 0.6899675296729249 and parameters: {'lambda_l1': 0.0033245647295823137, 'lambda_l2': 2.7333945263234923e-06}. Best is trial 45 with value: 0.6899674025098815. regularization_factors, val_score: 0.689967: 40%|#### | 8/20 [00:07<00:12, 1.00s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001603 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663071 valid's binary_logloss: 0.690424 Early stopping, best iteration is: [79] train's binary_logloss: 0.667888 valid's binary_logloss: 0.689968 regularization_factors, val_score: 0.689967: 45%|####5 | 9/20 [00:08<00:10, 1.01it/s][I 2020-09-27 04:57:57,415] Trial 51 finished with value: 0.6899675231589405 and parameters: {'lambda_l1': 0.0034231822486881607, 'lambda_l2': 2.5597586648530185e-06}. Best is trial 45 with value: 0.6899674025098815. regularization_factors, val_score: 0.689967: 45%|####5 | 9/20 [00:08<00:10, 1.01it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001568 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663071 valid's binary_logloss: 0.690424 Early stopping, best iteration is: [79] train's binary_logloss: 0.667888 valid's binary_logloss: 0.689968 regularization_factors, val_score: 0.689967: 50%|##### | 10/20 [00:09<00:09, 1.02it/s][I 2020-09-27 04:57:58,387] Trial 52 finished with value: 0.6899675209429271 and parameters: {'lambda_l1': 0.0034569782502191274, 'lambda_l2': 2.163106922428234e-06}. Best is trial 45 with value: 0.6899674025098815. regularization_factors, val_score: 0.689967: 50%|##### | 10/20 [00:09<00:09, 1.02it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001571 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663071 valid's binary_logloss: 0.690424 Early stopping, best iteration is: [79] train's binary_logloss: 0.667889 valid's binary_logloss: 0.689968 regularization_factors, val_score: 0.689967: 55%|#####5 | 11/20 [00:10<00:08, 1.02it/s][I 2020-09-27 04:57:59,369] Trial 53 finished with value: 0.6899675043589648 and parameters: {'lambda_l1': 0.003707901920209718, 'lambda_l2': 1.70752737693983e-06}. Best is trial 45 with value: 0.6899674025098815. regularization_factors, val_score: 0.689967: 55%|#####5 | 11/20 [00:10<00:08, 1.02it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001536 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.66306 valid's binary_logloss: 0.690447 Early stopping, best iteration is: [79] train's binary_logloss: 0.667889 valid's binary_logloss: 0.689967 regularization_factors, val_score: 0.689967: 60%|###### | 12/20 [00:11<00:07, 1.01it/s][I 2020-09-27 04:58:00,373] Trial 54 finished with value: 0.6899674606296653 and parameters: {'lambda_l1': 0.004368898887817153, 'lambda_l2': 1.2458711221699426e-06}. Best is trial 45 with value: 0.6899674025098815. regularization_factors, val_score: 0.689967: 60%|###### | 12/20 [00:11<00:07, 1.01it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001596 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.662937 valid's binary_logloss: 0.690557 Early stopping, best iteration is: [47] train's binary_logloss: 0.675852 valid's binary_logloss: 0.690271 regularization_factors, val_score: 0.689967: 65%|######5 | 13/20 [00:12<00:06, 1.05it/s][I 2020-09-27 04:58:01,228] Trial 55 finished with value: 0.6902710837967893 and parameters: {'lambda_l1': 0.07746158051498105, 'lambda_l2': 3.231480194675918e-08}. Best is trial 45 with value: 0.6899674025098815. regularization_factors, val_score: 0.689967: 65%|######5 | 13/20 [00:12<00:06, 1.05it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008971 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664683 valid's binary_logloss: 0.690143 Early stopping, best iteration is: [52] train's binary_logloss: 0.675081 valid's binary_logloss: 0.689647 regularization_factors, val_score: 0.689647: 70%|####### | 14/20 [00:13<00:05, 1.08it/s][I 2020-09-27 04:58:02,088] Trial 56 finished with value: 0.6896473829748281 and parameters: {'lambda_l1': 5.946750740071561e-06, 'lambda_l2': 2.6997857156008718}. Best is trial 56 with value: 0.6896473829748281. regularization_factors, val_score: 0.689647: 70%|####### | 14/20 [00:13<00:05, 1.08it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001490 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665021 valid's binary_logloss: 0.691158 Early stopping, best iteration is: [47] train's binary_logloss: 0.676401 valid's binary_logloss: 0.6902 regularization_factors, val_score: 0.689647: 75%|#######5 | 15/20 [00:14<00:04, 1.11it/s][I 2020-09-27 04:58:02,943] Trial 57 finished with value: 0.6902003315084454 and parameters: {'lambda_l1': 5.758495952716345e-07, 'lambda_l2': 3.755109192731262}. Best is trial 56 with value: 0.6896473829748281. regularization_factors, val_score: 0.689647: 75%|#######5 | 15/20 [00:14<00:04, 1.11it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001569 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663022 valid's binary_logloss: 0.690361 Early stopping, best iteration is: [48] train's binary_logloss: 0.675429 valid's binary_logloss: 0.689984 regularization_factors, val_score: 0.689647: 80%|######## | 16/20 [00:15<00:03, 1.14it/s][I 2020-09-27 04:58:03,771] Trial 58 finished with value: 0.6899844790092887 and parameters: {'lambda_l1': 2.0582999452425637e-05, 'lambda_l2': 0.0005026756920447598}. Best is trial 56 with value: 0.6896473829748281. regularization_factors, val_score: 0.689647: 80%|######## | 16/20 [00:15<00:03, 1.14it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001565 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663021 valid's binary_logloss: 0.690361 Early stopping, best iteration is: [48] train's binary_logloss: 0.675428 valid's binary_logloss: 0.689984 regularization_factors, val_score: 0.689647: 85%|########5 | 17/20 [00:16<00:02, 1.16it/s][I 2020-09-27 04:58:04,597] Trial 59 finished with value: 0.6899844953569758 and parameters: {'lambda_l1': 6.0992831053207785e-05, 'lambda_l2': 4.252592969567293e-08}. Best is trial 56 with value: 0.6896473829748281. regularization_factors, val_score: 0.689647: 85%|########5 | 17/20 [00:16<00:02, 1.16it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001929 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663089 valid's binary_logloss: 0.691109 Early stopping, best iteration is: [43] train's binary_logloss: 0.676731 valid's binary_logloss: 0.690043 regularization_factors, val_score: 0.689647: 90%|######### | 18/20 [00:16<00:01, 1.17it/s][I 2020-09-27 04:58:05,432] Trial 60 finished with value: 0.6900427743024334 and parameters: {'lambda_l1': 0.268567393042738, 'lambda_l2': 0.0003824677053871511}. Best is trial 56 with value: 0.6896473829748281. regularization_factors, val_score: 0.689647: 90%|######### | 18/20 [00:16<00:01, 1.17it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001538 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663047 valid's binary_logloss: 0.690658 Early stopping, best iteration is: [48] train's binary_logloss: 0.675429 valid's binary_logloss: 0.689984 regularization_factors, val_score: 0.689647: 95%|#########5| 19/20 [00:17<00:00, 1.17it/s][I 2020-09-27 04:58:06,282] Trial 61 finished with value: 0.6899844746373079 and parameters: {'lambda_l1': 0.0005260392810081732, 'lambda_l2': 2.7967911031858336e-07}. Best is trial 56 with value: 0.6896473829748281. regularization_factors, val_score: 0.689647: 95%|#########5| 19/20 [00:17<00:00, 1.17it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001556 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663199 valid's binary_logloss: 0.690198 Early stopping, best iteration is: [52] train's binary_logloss: 0.674513 valid's binary_logloss: 0.68975 regularization_factors, val_score: 0.689647: 100%|##########| 20/20 [00:18<00:00, 1.15it/s][I 2020-09-27 04:58:07,180] Trial 62 finished with value: 0.6897502601908707 and parameters: {'lambda_l1': 0.03166566393209921, 'lambda_l2': 3.6505882507400447e-05}. Best is trial 56 with value: 0.6896473829748281. regularization_factors, val_score: 0.689647: 100%|##########| 20/20 [00:18<00:00, 1.08it/s] min_data_in_leaf, val_score: 0.689647: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011144 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664472 valid's binary_logloss: 0.690172 [200] train's binary_logloss: 0.645537 valid's binary_logloss: 0.690997 Early stopping, best iteration is: [108] train's binary_logloss: 0.662894 valid's binary_logloss: 0.690057 min_data_in_leaf, val_score: 0.689647: 20%|## | 1/5 [00:01<00:04, 1.19s/it][I 2020-09-27 04:58:08,386] Trial 63 finished with value: 0.6900565147753162 and parameters: {'min_child_samples': 10}. Best is trial 63 with value: 0.6900565147753162. min_data_in_leaf, val_score: 0.689647: 20%|## | 1/5 [00:01<00:04, 1.19s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002284 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665148 valid's binary_logloss: 0.691028 Early stopping, best iteration is: [49] train's binary_logloss: 0.676081 valid's binary_logloss: 0.690354 min_data_in_leaf, val_score: 0.689647: 40%|#### | 2/5 [00:02<00:03, 1.10s/it][I 2020-09-27 04:58:09,276] Trial 64 finished with value: 0.6903537539012076 and parameters: {'min_child_samples': 100}. Best is trial 63 with value: 0.6900565147753162. min_data_in_leaf, val_score: 0.689647: 40%|#### | 2/5 [00:02<00:03, 1.10s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007936 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664812 valid's binary_logloss: 0.690985 Early stopping, best iteration is: [41] train's binary_logloss: 0.677793 valid's binary_logloss: 0.690549 min_data_in_leaf, val_score: 0.689647: 60%|###### | 3/5 [00:02<00:02, 1.02s/it][I 2020-09-27 04:58:10,108] Trial 65 finished with value: 0.690549034445801 and parameters: {'min_child_samples': 5}. Best is trial 63 with value: 0.6900565147753162. min_data_in_leaf, val_score: 0.689647: 60%|###### | 3/5 [00:02<00:02, 1.02s/it][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001560 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664729 valid's binary_logloss: 0.690892 Early stopping, best iteration is: [41] train's binary_logloss: 0.677757 valid's binary_logloss: 0.690151 min_data_in_leaf, val_score: 0.689647: 80%|######## | 4/5 [00:03<00:00, 1.03it/s][I 2020-09-27 04:58:10,959] Trial 66 finished with value: 0.6901508243301284 and parameters: {'min_child_samples': 25}. Best is trial 63 with value: 0.6900565147753162. min_data_in_leaf, val_score: 0.689647: 80%|######## | 4/5 [00:03<00:00, 1.03it/s][LightGBM] [Info] Number of positive: 46417, number of negative: 46609 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001581 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498968 -> initscore=-0.004128 [LightGBM] [Info] Start training from score -0.004128 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664643 valid's binary_logloss: 0.691412 Early stopping, best iteration is: [39] train's binary_logloss: 0.678314 valid's binary_logloss: 0.690208 min_data_in_leaf, val_score: 0.689647: 100%|##########| 5/5 [00:04<00:00, 1.07it/s][I 2020-09-27 04:58:11,803] Trial 67 finished with value: 0.6902082945895022 and parameters: {'min_child_samples': 50}. Best is trial 63 with value: 0.6900565147753162. min_data_in_leaf, val_score: 0.689647: 100%|##########| 5/5 [00:04<00:00, 1.08it/s] Fold : 6 [I 2020-09-27 04:58:11,870] A new study created in memory with name: no-name-dcbdec6e-2ceb-4d1e-86b5-bfbd5414d073 feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000995 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665269 valid's binary_logloss: 0.688894 [200] train's binary_logloss: 0.645599 valid's binary_logloss: 0.690352 Early stopping, best iteration is: [106] train's binary_logloss: 0.663937 valid's binary_logloss: 0.688771 feature_fraction, val_score: 0.688771: 14%|#4 | 1/7 [00:01<00:07, 1.19s/it][I 2020-09-27 04:58:13,071] Trial 0 finished with value: 0.6887706335922763 and parameters: {'feature_fraction': 0.5}. Best is trial 0 with value: 0.6887706335922763. feature_fraction, val_score: 0.688771: 14%|#4 | 1/7 [00:01<00:07, 1.19s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001614 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.662986 valid's binary_logloss: 0.689796 Early stopping, best iteration is: [32] train's binary_logloss: 0.679798 valid's binary_logloss: 0.689557 feature_fraction, val_score: 0.688771: 29%|##8 | 2/7 [00:01<00:05, 1.06s/it][I 2020-09-27 04:58:13,811] Trial 1 finished with value: 0.689557320235649 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 0 with value: 0.6887706335922763. feature_fraction, val_score: 0.688771: 29%|##8 | 2/7 [00:01<00:05, 1.06s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.015363 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663698 valid's binary_logloss: 0.689777 Early stopping, best iteration is: [62] train's binary_logloss: 0.672409 valid's binary_logloss: 0.689451 feature_fraction, val_score: 0.688771: 43%|####2 | 3/7 [00:02<00:03, 1.02it/s][I 2020-09-27 04:58:14,610] Trial 2 finished with value: 0.6894514131827376 and parameters: {'feature_fraction': 0.7}. Best is trial 0 with value: 0.6887706335922763. feature_fraction, val_score: 0.688771: 43%|####2 | 3/7 [00:02<00:03, 1.02it/s][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007904 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.66431 valid's binary_logloss: 0.689418 Early stopping, best iteration is: [82] train's binary_logloss: 0.668145 valid's binary_logloss: 0.689215 feature_fraction, val_score: 0.688771: 57%|#####7 | 4/7 [00:03<00:02, 1.07it/s][I 2020-09-27 04:58:15,452] Trial 3 finished with value: 0.6892151532007572 and parameters: {'feature_fraction': 0.6}. Best is trial 0 with value: 0.6887706335922763. feature_fraction, val_score: 0.688771: 57%|#####7 | 4/7 [00:03<00:02, 1.07it/s][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008838 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663214 valid's binary_logloss: 0.689646 Early stopping, best iteration is: [61] train's binary_logloss: 0.672363 valid's binary_logloss: 0.689118 feature_fraction, val_score: 0.688771: 71%|#######1 | 5/7 [00:04<00:01, 1.11it/s][I 2020-09-27 04:58:16,267] Trial 4 finished with value: 0.6891176758344856 and parameters: {'feature_fraction': 0.8}. Best is trial 0 with value: 0.6887706335922763. feature_fraction, val_score: 0.688771: 71%|#######1 | 5/7 [00:04<00:01, 1.11it/s][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000778 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.66613 valid's binary_logloss: 0.689486 Early stopping, best iteration is: [75] train's binary_logloss: 0.671175 valid's binary_logloss: 0.689183 feature_fraction, val_score: 0.688771: 86%|########5 | 6/7 [00:05<00:00, 1.13it/s][I 2020-09-27 04:58:17,113] Trial 5 finished with value: 0.6891830705136953 and parameters: {'feature_fraction': 0.4}. Best is trial 0 with value: 0.6887706335922763. feature_fraction, val_score: 0.688771: 86%|########5 | 6/7 [00:05<00:00, 1.13it/s][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001832 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.662507 valid's binary_logloss: 0.690168 Early stopping, best iteration is: [48] train's binary_logloss: 0.675204 valid's binary_logloss: 0.689566 feature_fraction, val_score: 0.688771: 100%|##########| 7/7 [00:07<00:00, 1.32s/it][I 2020-09-27 04:58:19,455] Trial 6 finished with value: 0.6895663490864646 and parameters: {'feature_fraction': 1.0}. Best is trial 0 with value: 0.6887706335922763. feature_fraction, val_score: 0.688771: 100%|##########| 7/7 [00:07<00:00, 1.08s/it] num_leaves, val_score: 0.688771: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000875 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.65635 valid's binary_logloss: 0.690173 Early stopping, best iteration is: [43] train's binary_logloss: 0.673607 valid's binary_logloss: 0.689502 num_leaves, val_score: 0.688771: 5%|5 | 1/20 [00:01<00:20, 1.08s/it][I 2020-09-27 04:58:20,546] Trial 7 finished with value: 0.6895015997019708 and parameters: {'num_leaves': 43}. Best is trial 7 with value: 0.6895015997019708. num_leaves, val_score: 0.688771: 5%|5 | 1/20 [00:01<00:20, 1.08s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000948 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.636388 valid's binary_logloss: 0.690447 Early stopping, best iteration is: [88] train's binary_logloss: 0.641912 valid's binary_logloss: 0.689924 num_leaves, val_score: 0.688771: 10%|# | 2/20 [00:02<00:20, 1.13s/it][I 2020-09-27 04:58:21,813] Trial 8 finished with value: 0.6899239348271381 and parameters: {'num_leaves': 73}. Best is trial 7 with value: 0.6895015997019708. num_leaves, val_score: 0.688771: 10%|# | 2/20 [00:02<00:20, 1.13s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001001 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.640263 valid's binary_logloss: 0.690357 Early stopping, best iteration is: [50] train's binary_logloss: 0.661952 valid's binary_logloss: 0.689585 num_leaves, val_score: 0.688771: 15%|#5 | 3/20 [00:03<00:18, 1.10s/it][I 2020-09-27 04:58:22,822] Trial 9 finished with value: 0.6895847174318764 and parameters: {'num_leaves': 67}. Best is trial 7 with value: 0.6895015997019708. num_leaves, val_score: 0.688771: 15%|#5 | 3/20 [00:03<00:18, 1.10s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001688 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.552597 valid's binary_logloss: 0.69652 Early stopping, best iteration is: [17] train's binary_logloss: 0.65881 valid's binary_logloss: 0.692576 num_leaves, val_score: 0.688771: 20%|## | 4/20 [00:04<00:19, 1.19s/it][I 2020-09-27 04:58:24,227] Trial 10 finished with value: 0.6925759831330955 and parameters: {'num_leaves': 228}. Best is trial 7 with value: 0.6895015997019708. num_leaves, val_score: 0.688771: 20%|## | 4/20 [00:04<00:19, 1.19s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001533 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.538026 valid's binary_logloss: 0.69752 Early stopping, best iteration is: [15] train's binary_logloss: 0.659639 valid's binary_logloss: 0.691473 num_leaves, val_score: 0.688771: 25%|##5 | 5/20 [00:06<00:18, 1.25s/it][I 2020-09-27 04:58:25,631] Trial 11 finished with value: 0.6914732359891139 and parameters: {'num_leaves': 254}. Best is trial 7 with value: 0.6895015997019708. num_leaves, val_score: 0.688771: 25%|##5 | 5/20 [00:06<00:18, 1.25s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001108 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.578766 valid's binary_logloss: 0.694226 Early stopping, best iteration is: [25] train's binary_logloss: 0.654454 valid's binary_logloss: 0.691199 num_leaves, val_score: 0.688771: 30%|### | 6/20 [00:07<00:17, 1.22s/it][I 2020-09-27 04:58:26,778] Trial 12 finished with value: 0.6911992488401045 and parameters: {'num_leaves': 174}. Best is trial 7 with value: 0.6895015997019708. num_leaves, val_score: 0.688771: 30%|### | 6/20 [00:07<00:17, 1.22s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000950 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.587943 valid's binary_logloss: 0.69436 Early stopping, best iteration is: [64] train's binary_logloss: 0.61787 valid's binary_logloss: 0.691342 num_leaves, val_score: 0.688771: 35%|###5 | 7/20 [00:08<00:16, 1.27s/it][I 2020-09-27 04:58:28,162] Trial 13 finished with value: 0.6913417018956572 and parameters: {'num_leaves': 156}. Best is trial 7 with value: 0.6895015997019708. num_leaves, val_score: 0.688771: 35%|###5 | 7/20 [00:08<00:16, 1.27s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001185 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.675867 valid's binary_logloss: 0.689217 [200] train's binary_logloss: 0.664751 valid's binary_logloss: 0.689817 Early stopping, best iteration is: [111] train's binary_logloss: 0.674533 valid's binary_logloss: 0.68909 num_leaves, val_score: 0.688771: 40%|#### | 8/20 [00:09<00:14, 1.18s/it][I 2020-09-27 04:58:29,127] Trial 14 finished with value: 0.6890903374716104 and parameters: {'num_leaves': 17}. Best is trial 14 with value: 0.6890903374716104. num_leaves, val_score: 0.688771: 40%|#### | 8/20 [00:09<00:14, 1.18s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001614 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688527 valid's binary_logloss: 0.690133 [200] train's binary_logloss: 0.686734 valid's binary_logloss: 0.689033 [300] train's binary_logloss: 0.685468 valid's binary_logloss: 0.688641 [400] train's binary_logloss: 0.684369 valid's binary_logloss: 0.688513 Early stopping, best iteration is: [375] train's binary_logloss: 0.684633 valid's binary_logloss: 0.68848 num_leaves, val_score: 0.688480: 45%|####5 | 9/20 [00:11<00:14, 1.32s/it][I 2020-09-27 04:58:30,769] Trial 15 finished with value: 0.6884795876194431 and parameters: {'num_leaves': 3}. Best is trial 15 with value: 0.6884795876194431. num_leaves, val_score: 0.688480: 45%|####5 | 9/20 [00:11<00:14, 1.32s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001058 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.610458 valid's binary_logloss: 0.69108 Early stopping, best iteration is: [49] train's binary_logloss: 0.645426 valid's binary_logloss: 0.690053 num_leaves, val_score: 0.688480: 50%|##### | 10/20 [00:12<00:12, 1.27s/it][I 2020-09-27 04:58:31,926] Trial 16 finished with value: 0.6900527437374864 and parameters: {'num_leaves': 115}. Best is trial 15 with value: 0.6884795876194431. num_leaves, val_score: 0.688480: 50%|##### | 10/20 [00:12<00:12, 1.27s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000997 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688527 valid's binary_logloss: 0.690133 [200] train's binary_logloss: 0.686734 valid's binary_logloss: 0.689033 [300] train's binary_logloss: 0.685468 valid's binary_logloss: 0.688641 [400] train's binary_logloss: 0.684369 valid's binary_logloss: 0.688513 Early stopping, best iteration is: [375] train's binary_logloss: 0.684633 valid's binary_logloss: 0.68848 num_leaves, val_score: 0.688480: 55%|#####5 | 11/20 [00:14<00:12, 1.37s/it][I 2020-09-27 04:58:33,539] Trial 17 finished with value: 0.6884795876194431 and parameters: {'num_leaves': 3}. Best is trial 15 with value: 0.6884795876194431. num_leaves, val_score: 0.688480: 55%|#####5 | 11/20 [00:14<00:12, 1.37s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000988 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.690025 valid's binary_logloss: 0.691003 [200] train's binary_logloss: 0.688901 valid's binary_logloss: 0.69002 [300] train's binary_logloss: 0.688185 valid's binary_logloss: 0.68941 [400] train's binary_logloss: 0.687683 valid's binary_logloss: 0.688943 [500] train's binary_logloss: 0.687304 valid's binary_logloss: 0.688654 [600] train's binary_logloss: 0.687004 valid's binary_logloss: 0.688502 [700] train's binary_logloss: 0.686761 valid's binary_logloss: 0.688428 [800] train's binary_logloss: 0.686556 valid's binary_logloss: 0.688335 [900] train's binary_logloss: 0.68638 valid's binary_logloss: 0.688325 [1000] train's binary_logloss: 0.686227 valid's binary_logloss: 0.688255 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.686227 valid's binary_logloss: 0.688255 num_leaves, val_score: 0.688255: 60%|###### | 12/20 [00:17<00:15, 1.92s/it][I 2020-09-27 04:58:36,728] Trial 18 finished with value: 0.6882551844633495 and parameters: {'num_leaves': 2}. Best is trial 18 with value: 0.6882551844633495. num_leaves, val_score: 0.688255: 60%|###### | 12/20 [00:17<00:15, 1.92s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001002 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.683515 valid's binary_logloss: 0.689389 [200] train's binary_logloss: 0.678306 valid's binary_logloss: 0.689315 [300] train's binary_logloss: 0.673658 valid's binary_logloss: 0.688984 [400] train's binary_logloss: 0.669291 valid's binary_logloss: 0.689314 Early stopping, best iteration is: [317] train's binary_logloss: 0.6729 valid's binary_logloss: 0.688929 num_leaves, val_score: 0.688255: 65%|######5 | 13/20 [00:18<00:12, 1.81s/it][I 2020-09-27 04:58:38,296] Trial 19 finished with value: 0.68892857284493 and parameters: {'num_leaves': 8}. Best is trial 18 with value: 0.6882551844633495. num_leaves, val_score: 0.688255: 65%|######5 | 13/20 [00:18<00:12, 1.81s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000915 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.62055 valid's binary_logloss: 0.692044 Early stopping, best iteration is: [29] train's binary_logloss: 0.66542 valid's binary_logloss: 0.690281 num_leaves, val_score: 0.688255: 70%|####### | 14/20 [00:19<00:09, 1.55s/it][I 2020-09-27 04:58:39,243] Trial 20 finished with value: 0.6902812591323769 and parameters: {'num_leaves': 99}. Best is trial 18 with value: 0.6882551844633495. num_leaves, val_score: 0.688255: 70%|####### | 14/20 [00:19<00:09, 1.55s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001482 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.684441 valid's binary_logloss: 0.689008 [200] train's binary_logloss: 0.679968 valid's binary_logloss: 0.688832 [300] train's binary_logloss: 0.676016 valid's binary_logloss: 0.688309 [400] train's binary_logloss: 0.672203 valid's binary_logloss: 0.688492 Early stopping, best iteration is: [301] train's binary_logloss: 0.675968 valid's binary_logloss: 0.688303 num_leaves, val_score: 0.688255: 75%|#######5 | 15/20 [00:21<00:07, 1.52s/it][I 2020-09-27 04:58:40,689] Trial 21 finished with value: 0.6883028732185514 and parameters: {'num_leaves': 7}. Best is trial 18 with value: 0.6882551844633495. num_leaves, val_score: 0.688255: 75%|#######5 | 15/20 [00:21<00:07, 1.52s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000968 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.662251 valid's binary_logloss: 0.689855 Early stopping, best iteration is: [71] train's binary_logloss: 0.669171 valid's binary_logloss: 0.689288 num_leaves, val_score: 0.688255: 80%|######## | 16/20 [00:22<00:05, 1.33s/it][I 2020-09-27 04:58:41,585] Trial 22 finished with value: 0.6892878269668582 and parameters: {'num_leaves': 35}. Best is trial 18 with value: 0.6882551844633495. num_leaves, val_score: 0.688255: 80%|######## | 16/20 [00:22<00:05, 1.33s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000967 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68536 valid's binary_logloss: 0.689552 [200] train's binary_logloss: 0.681488 valid's binary_logloss: 0.689209 [300] train's binary_logloss: 0.678139 valid's binary_logloss: 0.68914 Early stopping, best iteration is: [279] train's binary_logloss: 0.67879 valid's binary_logloss: 0.689 num_leaves, val_score: 0.688255: 85%|########5 | 17/20 [00:23<00:04, 1.35s/it][I 2020-09-27 04:58:42,989] Trial 23 finished with value: 0.6889999863028395 and parameters: {'num_leaves': 6}. Best is trial 18 with value: 0.6882551844633495. num_leaves, val_score: 0.688255: 85%|########5 | 17/20 [00:23<00:04, 1.35s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000969 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.653975 valid's binary_logloss: 0.689905 Early stopping, best iteration is: [63] train's binary_logloss: 0.665486 valid's binary_logloss: 0.689503 num_leaves, val_score: 0.688255: 90%|######### | 18/20 [00:24<00:02, 1.25s/it][I 2020-09-27 04:58:43,991] Trial 24 finished with value: 0.6895028538395134 and parameters: {'num_leaves': 47}. Best is trial 18 with value: 0.6882551844633495. num_leaves, val_score: 0.688255: 90%|######### | 18/20 [00:24<00:02, 1.25s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009425 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688527 valid's binary_logloss: 0.690133 [200] train's binary_logloss: 0.686734 valid's binary_logloss: 0.689033 [300] train's binary_logloss: 0.685468 valid's binary_logloss: 0.688641 [400] train's binary_logloss: 0.684369 valid's binary_logloss: 0.688513 Early stopping, best iteration is: [375] train's binary_logloss: 0.684633 valid's binary_logloss: 0.68848 num_leaves, val_score: 0.688255: 95%|#########5| 19/20 [00:26<00:01, 1.34s/it][I 2020-09-27 04:58:45,542] Trial 25 finished with value: 0.6884795876194431 and parameters: {'num_leaves': 3}. Best is trial 18 with value: 0.6882551844633495. num_leaves, val_score: 0.688255: 95%|#########5| 19/20 [00:26<00:01, 1.34s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000942 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.634107 valid's binary_logloss: 0.691117 Early stopping, best iteration is: [34] train's binary_logloss: 0.66756 valid's binary_logloss: 0.689318 num_leaves, val_score: 0.688255: 100%|##########| 20/20 [00:26<00:00, 1.21s/it][I 2020-09-27 04:58:46,434] Trial 26 finished with value: 0.6893183889206451 and parameters: {'num_leaves': 76}. Best is trial 18 with value: 0.6882551844633495. num_leaves, val_score: 0.688255: 100%|##########| 20/20 [00:26<00:00, 1.35s/it] bagging, val_score: 0.688255: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000882 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577 [200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588 [300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856 [400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313 [500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084 [600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043 [700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014 Early stopping, best iteration is: [671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857 bagging, val_score: 0.687857: 10%|# | 1/10 [00:02<00:23, 2.59s/it][I 2020-09-27 04:58:49,036] Trial 27 finished with value: 0.6878571684251173 and parameters: {'bagging_fraction': 0.7285342265331846, 'bagging_freq': 5}. Best is trial 27 with value: 0.6878571684251173. bagging, val_score: 0.687857: 10%|# | 1/10 [00:02<00:23, 2.59s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000976 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689824 valid's binary_logloss: 0.690686 [200] train's binary_logloss: 0.688555 valid's binary_logloss: 0.68967 [300] train's binary_logloss: 0.68776 valid's binary_logloss: 0.688991 [400] train's binary_logloss: 0.687201 valid's binary_logloss: 0.688421 [500] train's binary_logloss: 0.68677 valid's binary_logloss: 0.688248 [600] train's binary_logloss: 0.686428 valid's binary_logloss: 0.688112 [700] train's binary_logloss: 0.686154 valid's binary_logloss: 0.688132 Early stopping, best iteration is: [670] train's binary_logloss: 0.686227 valid's binary_logloss: 0.688037 bagging, val_score: 0.687857: 20%|## | 2/10 [00:05<00:20, 2.62s/it][I 2020-09-27 04:58:51,737] Trial 28 finished with value: 0.6880373748785599 and parameters: {'bagging_fraction': 0.7377723938528014, 'bagging_freq': 5}. Best is trial 27 with value: 0.6878571684251173. bagging, val_score: 0.687857: 20%|## | 2/10 [00:05<00:20, 2.62s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001053 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689798 valid's binary_logloss: 0.690553 [200] train's binary_logloss: 0.688509 valid's binary_logloss: 0.689611 [300] train's binary_logloss: 0.687723 valid's binary_logloss: 0.688922 [400] train's binary_logloss: 0.68716 valid's binary_logloss: 0.688433 [500] train's binary_logloss: 0.686722 valid's binary_logloss: 0.688211 [600] train's binary_logloss: 0.686375 valid's binary_logloss: 0.688204 Early stopping, best iteration is: [555] train's binary_logloss: 0.686527 valid's binary_logloss: 0.688082 bagging, val_score: 0.687857: 30%|### | 3/10 [00:07<00:17, 2.49s/it][I 2020-09-27 04:58:53,937] Trial 29 finished with value: 0.6880818176819911 and parameters: {'bagging_fraction': 0.7152227689225483, 'bagging_freq': 5}. Best is trial 27 with value: 0.6878571684251173. bagging, val_score: 0.687857: 30%|### | 3/10 [00:07<00:17, 2.49s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000973 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689833 valid's binary_logloss: 0.690598 [200] train's binary_logloss: 0.688544 valid's binary_logloss: 0.68966 [300] train's binary_logloss: 0.687749 valid's binary_logloss: 0.688918 [400] train's binary_logloss: 0.687185 valid's binary_logloss: 0.688393 [500] train's binary_logloss: 0.686741 valid's binary_logloss: 0.688141 [600] train's binary_logloss: 0.686404 valid's binary_logloss: 0.688102 Early stopping, best iteration is: [555] train's binary_logloss: 0.686547 valid's binary_logloss: 0.688023 bagging, val_score: 0.687857: 40%|#### | 4/10 [00:09<00:14, 2.43s/it][I 2020-09-27 04:58:56,211] Trial 30 finished with value: 0.6880233595037442 and parameters: {'bagging_fraction': 0.7356006552369923, 'bagging_freq': 5}. Best is trial 27 with value: 0.6878571684251173. bagging, val_score: 0.687857: 40%|#### | 4/10 [00:09<00:14, 2.43s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000978 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689826 valid's binary_logloss: 0.690712 [200] train's binary_logloss: 0.688569 valid's binary_logloss: 0.689662 [300] train's binary_logloss: 0.687746 valid's binary_logloss: 0.688891 [400] train's binary_logloss: 0.687195 valid's binary_logloss: 0.688435 [500] train's binary_logloss: 0.686757 valid's binary_logloss: 0.688266 [600] train's binary_logloss: 0.686422 valid's binary_logloss: 0.688159 Early stopping, best iteration is: [552] train's binary_logloss: 0.686576 valid's binary_logloss: 0.688066 bagging, val_score: 0.687857: 50%|##### | 5/10 [00:12<00:11, 2.37s/it][I 2020-09-27 04:58:58,459] Trial 31 finished with value: 0.6880661932157487 and parameters: {'bagging_fraction': 0.7317912944017423, 'bagging_freq': 5}. Best is trial 27 with value: 0.6878571684251173. bagging, val_score: 0.687857: 50%|##### | 5/10 [00:12<00:11, 2.37s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001775 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689822 valid's binary_logloss: 0.690641 [200] train's binary_logloss: 0.688546 valid's binary_logloss: 0.689616 [300] train's binary_logloss: 0.687753 valid's binary_logloss: 0.688948 [400] train's binary_logloss: 0.687196 valid's binary_logloss: 0.68841 [500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688198 [600] train's binary_logloss: 0.686419 valid's binary_logloss: 0.688131 Early stopping, best iteration is: [555] train's binary_logloss: 0.686564 valid's binary_logloss: 0.688079 bagging, val_score: 0.687857: 60%|###### | 6/10 [00:14<00:09, 2.34s/it][I 2020-09-27 04:59:00,714] Trial 32 finished with value: 0.6880787095164863 and parameters: {'bagging_fraction': 0.7367657281985888, 'bagging_freq': 5}. Best is trial 27 with value: 0.6878571684251173. bagging, val_score: 0.687857: 60%|###### | 6/10 [00:14<00:09, 2.34s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000997 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689824 valid's binary_logloss: 0.690686 [200] train's binary_logloss: 0.688555 valid's binary_logloss: 0.68967 [300] train's binary_logloss: 0.68776 valid's binary_logloss: 0.688991 [400] train's binary_logloss: 0.687201 valid's binary_logloss: 0.688421 [500] train's binary_logloss: 0.68677 valid's binary_logloss: 0.688248 [600] train's binary_logloss: 0.686428 valid's binary_logloss: 0.688112 [700] train's binary_logloss: 0.686154 valid's binary_logloss: 0.688132 Early stopping, best iteration is: [670] train's binary_logloss: 0.686227 valid's binary_logloss: 0.688037 bagging, val_score: 0.687857: 70%|####### | 7/10 [00:16<00:07, 2.40s/it][I 2020-09-27 04:59:03,266] Trial 33 finished with value: 0.6880373748785599 and parameters: {'bagging_fraction': 0.7377865570355829, 'bagging_freq': 5}. Best is trial 27 with value: 0.6878571684251173. bagging, val_score: 0.687857: 70%|####### | 7/10 [00:16<00:07, 2.40s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000994 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689832 valid's binary_logloss: 0.69068 [200] train's binary_logloss: 0.688571 valid's binary_logloss: 0.689603 [300] train's binary_logloss: 0.687772 valid's binary_logloss: 0.688883 [400] train's binary_logloss: 0.687221 valid's binary_logloss: 0.68836 [500] train's binary_logloss: 0.686782 valid's binary_logloss: 0.688152 [600] train's binary_logloss: 0.686435 valid's binary_logloss: 0.688108 Early stopping, best iteration is: [555] train's binary_logloss: 0.686576 valid's binary_logloss: 0.687995 bagging, val_score: 0.687857: 80%|######## | 8/10 [00:19<00:04, 2.35s/it][I 2020-09-27 04:59:05,479] Trial 34 finished with value: 0.6879951969996739 and parameters: {'bagging_fraction': 0.7515162282300969, 'bagging_freq': 5}. Best is trial 27 with value: 0.6878571684251173. bagging, val_score: 0.687857: 80%|######## | 8/10 [00:19<00:04, 2.35s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001010 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689966 valid's binary_logloss: 0.690938 [200] train's binary_logloss: 0.688773 valid's binary_logloss: 0.689819 [300] train's binary_logloss: 0.688013 valid's binary_logloss: 0.689196 [400] train's binary_logloss: 0.687467 valid's binary_logloss: 0.688806 [500] train's binary_logloss: 0.687057 valid's binary_logloss: 0.688497 [600] train's binary_logloss: 0.686739 valid's binary_logloss: 0.688298 [700] train's binary_logloss: 0.686478 valid's binary_logloss: 0.688315 [800] train's binary_logloss: 0.686255 valid's binary_logloss: 0.688233 [900] train's binary_logloss: 0.686067 valid's binary_logloss: 0.688251 Early stopping, best iteration is: [869] train's binary_logloss: 0.686123 valid's binary_logloss: 0.688191 bagging, val_score: 0.687857: 90%|######### | 9/10 [00:22<00:02, 2.68s/it][I 2020-09-27 04:59:08,934] Trial 35 finished with value: 0.6881912997605557 and parameters: {'bagging_fraction': 0.9421235215974794, 'bagging_freq': 4}. Best is trial 27 with value: 0.6878571684251173. bagging, val_score: 0.687857: 90%|######### | 9/10 [00:22<00:02, 2.68s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001114 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689917 valid's binary_logloss: 0.690846 [200] train's binary_logloss: 0.688683 valid's binary_logloss: 0.689724 [300] train's binary_logloss: 0.68788 valid's binary_logloss: 0.689127 [400] train's binary_logloss: 0.687336 valid's binary_logloss: 0.6886 [500] train's binary_logloss: 0.686912 valid's binary_logloss: 0.688322 [600] train's binary_logloss: 0.686585 valid's binary_logloss: 0.688191 [700] train's binary_logloss: 0.686301 valid's binary_logloss: 0.688109 [800] train's binary_logloss: 0.686068 valid's binary_logloss: 0.688108 [900] train's binary_logloss: 0.685866 valid's binary_logloss: 0.688065 Early stopping, best iteration is: [869] train's binary_logloss: 0.68593 valid's binary_logloss: 0.687939 bagging, val_score: 0.687857: 100%|##########| 10/10 [00:25<00:00, 2.86s/it][I 2020-09-27 04:59:12,229] Trial 36 finished with value: 0.6879386559102665 and parameters: {'bagging_fraction': 0.8372483508530265, 'bagging_freq': 7}. Best is trial 27 with value: 0.6878571684251173. bagging, val_score: 0.687857: 100%|##########| 10/10 [00:25<00:00, 2.58s/it] feature_fraction_stage2, val_score: 0.687857: 0%| | 0/6 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001013 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577 [200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588 [300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856 [400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313 [500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084 [600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043 [700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014 Early stopping, best iteration is: [671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857 feature_fraction_stage2, val_score: 0.687857: 17%|#6 | 1/6 [00:02<00:12, 2.55s/it][I 2020-09-27 04:59:14,795] Trial 37 finished with value: 0.6878571684251175 and parameters: {'feature_fraction': 0.484}. Best is trial 37 with value: 0.6878571684251175. feature_fraction_stage2, val_score: 0.687857: 17%|#6 | 1/6 [00:02<00:12, 2.55s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000890 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690634 [200] train's binary_logloss: 0.688568 valid's binary_logloss: 0.689696 [300] train's binary_logloss: 0.687761 valid's binary_logloss: 0.688942 [400] train's binary_logloss: 0.687217 valid's binary_logloss: 0.688449 [500] train's binary_logloss: 0.686776 valid's binary_logloss: 0.688296 [600] train's binary_logloss: 0.686427 valid's binary_logloss: 0.688198 Early stopping, best iteration is: [552] train's binary_logloss: 0.686591 valid's binary_logloss: 0.688123 feature_fraction_stage2, val_score: 0.687857: 33%|###3 | 2/6 [00:04<00:09, 2.42s/it][I 2020-09-27 04:59:16,923] Trial 38 finished with value: 0.6881228060729746 and parameters: {'feature_fraction': 0.45199999999999996}. Best is trial 37 with value: 0.6878571684251175. feature_fraction_stage2, val_score: 0.687857: 33%|###3 | 2/6 [00:04<00:09, 2.42s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000977 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577 [200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588 [300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856 [400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313 [500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084 [600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043 [700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014 Early stopping, best iteration is: [671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857 feature_fraction_stage2, val_score: 0.687857: 50%|##### | 3/6 [00:07<00:07, 2.47s/it][I 2020-09-27 04:59:19,495] Trial 39 finished with value: 0.6878571684251175 and parameters: {'feature_fraction': 0.516}. Best is trial 37 with value: 0.6878571684251175. feature_fraction_stage2, val_score: 0.687857: 50%|##### | 3/6 [00:07<00:07, 2.47s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000891 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689834 valid's binary_logloss: 0.690644 [200] train's binary_logloss: 0.688576 valid's binary_logloss: 0.689601 [300] train's binary_logloss: 0.687774 valid's binary_logloss: 0.688902 [400] train's binary_logloss: 0.687209 valid's binary_logloss: 0.688349 [500] train's binary_logloss: 0.686776 valid's binary_logloss: 0.688135 [600] train's binary_logloss: 0.68644 valid's binary_logloss: 0.688077 Early stopping, best iteration is: [555] train's binary_logloss: 0.686591 valid's binary_logloss: 0.687962 feature_fraction_stage2, val_score: 0.687857: 67%|######6 | 4/6 [00:09<00:04, 2.37s/it][I 2020-09-27 04:59:21,646] Trial 40 finished with value: 0.6879623345018646 and parameters: {'feature_fraction': 0.42}. Best is trial 37 with value: 0.6878571684251175. feature_fraction_stage2, val_score: 0.687857: 67%|######6 | 4/6 [00:09<00:04, 2.37s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001099 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689815 valid's binary_logloss: 0.690584 [200] train's binary_logloss: 0.688543 valid's binary_logloss: 0.689546 [300] train's binary_logloss: 0.687738 valid's binary_logloss: 0.688813 [400] train's binary_logloss: 0.687187 valid's binary_logloss: 0.688335 [500] train's binary_logloss: 0.686757 valid's binary_logloss: 0.688142 [600] train's binary_logloss: 0.686403 valid's binary_logloss: 0.688107 Early stopping, best iteration is: [555] train's binary_logloss: 0.68656 valid's binary_logloss: 0.688002 feature_fraction_stage2, val_score: 0.687857: 83%|########3 | 5/6 [00:11<00:02, 2.33s/it][I 2020-09-27 04:59:23,892] Trial 41 finished with value: 0.6880017676081877 and parameters: {'feature_fraction': 0.58}. Best is trial 37 with value: 0.6878571684251175. feature_fraction_stage2, val_score: 0.687857: 83%|########3 | 5/6 [00:11<00:02, 2.33s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001021 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689822 valid's binary_logloss: 0.690658 [200] train's binary_logloss: 0.688541 valid's binary_logloss: 0.689595 [300] train's binary_logloss: 0.687744 valid's binary_logloss: 0.688805 [400] train's binary_logloss: 0.687186 valid's binary_logloss: 0.688373 [500] train's binary_logloss: 0.686758 valid's binary_logloss: 0.688191 [600] train's binary_logloss: 0.68641 valid's binary_logloss: 0.688144 [700] train's binary_logloss: 0.686139 valid's binary_logloss: 0.688033 Early stopping, best iteration is: [671] train's binary_logloss: 0.686211 valid's binary_logloss: 0.687908 feature_fraction_stage2, val_score: 0.687857: 100%|##########| 6/6 [00:14<00:00, 2.41s/it][I 2020-09-27 04:59:26,490] Trial 42 finished with value: 0.6879081347102528 and parameters: {'feature_fraction': 0.5479999999999999}. Best is trial 37 with value: 0.6878571684251175. feature_fraction_stage2, val_score: 0.687857: 100%|##########| 6/6 [00:14<00:00, 2.38s/it] regularization_factors, val_score: 0.687857: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000956 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689827 valid's binary_logloss: 0.690607 [200] train's binary_logloss: 0.688566 valid's binary_logloss: 0.689648 [300] train's binary_logloss: 0.687757 valid's binary_logloss: 0.688831 [400] train's binary_logloss: 0.687199 valid's binary_logloss: 0.688357 [500] train's binary_logloss: 0.686766 valid's binary_logloss: 0.688172 [600] train's binary_logloss: 0.686415 valid's binary_logloss: 0.688061 Early stopping, best iteration is: [555] train's binary_logloss: 0.686571 valid's binary_logloss: 0.68802 regularization_factors, val_score: 0.687857: 5%|5 | 1/20 [00:02<00:41, 2.18s/it][I 2020-09-27 04:59:28,683] Trial 43 finished with value: 0.6880201782378496 and parameters: {'lambda_l1': 0.09717855595894546, 'lambda_l2': 2.599101232080336e-05}. Best is trial 43 with value: 0.6880201782378496. regularization_factors, val_score: 0.687857: 5%|5 | 1/20 [00:02<00:41, 2.18s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001017 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689864 valid's binary_logloss: 0.690597 [200] train's binary_logloss: 0.688608 valid's binary_logloss: 0.68959 [300] train's binary_logloss: 0.687806 valid's binary_logloss: 0.688794 [400] train's binary_logloss: 0.687261 valid's binary_logloss: 0.688287 [500] train's binary_logloss: 0.686828 valid's binary_logloss: 0.688121 [600] train's binary_logloss: 0.686479 valid's binary_logloss: 0.68802 Early stopping, best iteration is: [552] train's binary_logloss: 0.686646 valid's binary_logloss: 0.68794 regularization_factors, val_score: 0.687857: 10%|# | 2/20 [00:04<00:39, 2.19s/it][I 2020-09-27 04:59:30,884] Trial 44 finished with value: 0.6879401204038632 and parameters: {'lambda_l1': 2.27626769111772e-08, 'lambda_l2': 3.9029986997360395}. Best is trial 44 with value: 0.6879401204038632. regularization_factors, val_score: 0.687857: 10%|# | 2/20 [00:04<00:39, 2.19s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000975 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577 [200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588 [300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856 [400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313 [500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084 [600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043 [700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014 Early stopping, best iteration is: [671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857 regularization_factors, val_score: 0.687857: 15%|#5 | 3/20 [00:06<00:39, 2.30s/it][I 2020-09-27 04:59:33,437] Trial 45 finished with value: 0.6878571684254513 and parameters: {'lambda_l1': 6.262982988092863e-08, 'lambda_l2': 4.447887374533072e-08}. Best is trial 45 with value: 0.6878571684254513. regularization_factors, val_score: 0.687857: 15%|#5 | 3/20 [00:06<00:39, 2.30s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000939 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577 [200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588 [300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856 [400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313 [500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084 [600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043 [700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014 Early stopping, best iteration is: [671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857 regularization_factors, val_score: 0.687857: 20%|## | 4/20 [00:09<00:37, 2.37s/it][I 2020-09-27 04:59:35,984] Trial 46 finished with value: 0.6878571684251785 and parameters: {'lambda_l1': 1.0695178382125108e-08, 'lambda_l2': 1.4853383476880933e-08}. Best is trial 46 with value: 0.6878571684251785. regularization_factors, val_score: 0.687857: 20%|## | 4/20 [00:09<00:37, 2.37s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000993 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577 [200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588 [300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856 [400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313 [500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084 [600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043 [700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014 Early stopping, best iteration is: [671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857 regularization_factors, val_score: 0.687857: 25%|##5 | 5/20 [00:12<00:36, 2.43s/it][I 2020-09-27 04:59:38,551] Trial 47 finished with value: 0.6878571684252762 and parameters: {'lambda_l1': 1.8418017989423836e-08, 'lambda_l2': 1.1880874452318282e-08}. Best is trial 46 with value: 0.6878571684251785. regularization_factors, val_score: 0.687857: 25%|##5 | 5/20 [00:12<00:36, 2.43s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000984 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577 [200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588 [300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856 [400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313 [500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084 [600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043 [700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014 Early stopping, best iteration is: [671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857 regularization_factors, val_score: 0.687857: 30%|### | 6/20 [00:14<00:34, 2.48s/it][I 2020-09-27 04:59:41,131] Trial 48 finished with value: 0.687857168425301 and parameters: {'lambda_l1': 1.8864997222383136e-08, 'lambda_l2': 1.1856973045132024e-08}. Best is trial 46 with value: 0.6878571684251785. regularization_factors, val_score: 0.687857: 30%|### | 6/20 [00:14<00:34, 2.48s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000984 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577 [200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588 [300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856 [400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313 [500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084 [600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043 [700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014 Early stopping, best iteration is: [671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857 regularization_factors, val_score: 0.687857: 35%|###5 | 7/20 [00:17<00:32, 2.53s/it][I 2020-09-27 04:59:43,792] Trial 49 finished with value: 0.6878571685137861 and parameters: {'lambda_l1': 2.3008656218442958e-05, 'lambda_l2': 1.0076176570249505e-08}. Best is trial 46 with value: 0.6878571684251785. regularization_factors, val_score: 0.687857: 35%|###5 | 7/20 [00:17<00:32, 2.53s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000984 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577 [200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588 [300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856 [400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313 [500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084 [600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043 [700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014 Early stopping, best iteration is: [671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857 regularization_factors, val_score: 0.687857: 40%|#### | 8/20 [00:19<00:30, 2.54s/it][I 2020-09-27 04:59:46,353] Trial 50 finished with value: 0.6878571684048133 and parameters: {'lambda_l1': 1.0325498267148577e-05, 'lambda_l2': 9.965999410553949e-06}. Best is trial 50 with value: 0.6878571684048133. regularization_factors, val_score: 0.687857: 40%|#### | 8/20 [00:19<00:30, 2.54s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001025 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577 [200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588 [300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856 [400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313 [500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084 [600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043 [700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014 Early stopping, best iteration is: [671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857 regularization_factors, val_score: 0.687857: 45%|####5 | 9/20 [00:22<00:28, 2.55s/it][I 2020-09-27 04:59:48,931] Trial 51 finished with value: 0.6878571684428725 and parameters: {'lambda_l1': 1.7588444362019025e-05, 'lambda_l2': 8.259367220543736e-06}. Best is trial 50 with value: 0.6878571684048133. regularization_factors, val_score: 0.687857: 45%|####5 | 9/20 [00:22<00:28, 2.55s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000992 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577 [200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588 [300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856 [400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313 [500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084 [600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043 [700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014 Early stopping, best iteration is: [671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857 regularization_factors, val_score: 0.687857: 50%|##### | 10/20 [00:25<00:25, 2.57s/it][I 2020-09-27 04:59:51,534] Trial 52 finished with value: 0.6878571684244191 and parameters: {'lambda_l1': 5.589768247317659e-07, 'lambda_l2': 5.973648036426743e-07}. Best is trial 50 with value: 0.6878571684048133. regularization_factors, val_score: 0.687857: 50%|##### | 10/20 [00:25<00:25, 2.57s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000986 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577 [200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588 [300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856 [400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313 [500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084 [600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043 [700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014 Early stopping, best iteration is: [671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857 regularization_factors, val_score: 0.687857: 55%|#####5 | 11/20 [00:27<00:23, 2.56s/it][I 2020-09-27 04:59:54,086] Trial 53 finished with value: 0.6878571684026484 and parameters: {'lambda_l1': 1.4374237878755214e-06, 'lambda_l2': 4.663535296869216e-06}. Best is trial 53 with value: 0.6878571684026484. regularization_factors, val_score: 0.687857: 55%|#####5 | 11/20 [00:27<00:23, 2.56s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000982 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577 [200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588 [300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856 [400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313 [500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084 [600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043 [700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014 Early stopping, best iteration is: [671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857 regularization_factors, val_score: 0.687857: 60%|###### | 12/20 [00:30<00:20, 2.57s/it][I 2020-09-27 04:59:56,659] Trial 54 finished with value: 0.6878571684225362 and parameters: {'lambda_l1': 4.5280682776478785e-06, 'lambda_l2': 3.232830506287648e-06}. Best is trial 53 with value: 0.6878571684026484. regularization_factors, val_score: 0.687857: 60%|###### | 12/20 [00:30<00:20, 2.57s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001020 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577 [200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588 [300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856 [400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313 [500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084 [600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043 [700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014 Early stopping, best iteration is: [671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857 regularization_factors, val_score: 0.687857: 65%|######5 | 13/20 [00:32<00:17, 2.56s/it][I 2020-09-27 04:59:59,214] Trial 55 finished with value: 0.6878571684052713 and parameters: {'lambda_l1': 2.2260343593678247e-06, 'lambda_l2': 4.664732251084378e-06}. Best is trial 53 with value: 0.6878571684026484. regularization_factors, val_score: 0.687857: 65%|######5 | 13/20 [00:32<00:17, 2.56s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001004 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577 [200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588 [300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856 [400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313 [500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084 [600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043 [700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014 Early stopping, best iteration is: [671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857 regularization_factors, val_score: 0.687857: 70%|####### | 14/20 [00:35<00:15, 2.56s/it][I 2020-09-27 05:00:01,765] Trial 56 finished with value: 0.6878571684225463 and parameters: {'lambda_l1': 2.731664642116516e-06, 'lambda_l2': 2.242416999990266e-06}. Best is trial 53 with value: 0.6878571684026484. regularization_factors, val_score: 0.687857: 70%|####### | 14/20 [00:35<00:15, 2.56s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000988 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577 [200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588 [300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856 [400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313 [500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084 [600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043 [700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014 Early stopping, best iteration is: [671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857 regularization_factors, val_score: 0.687857: 75%|#######5 | 15/20 [00:37<00:12, 2.56s/it][I 2020-09-27 05:00:04,326] Trial 57 finished with value: 0.6878571684168346 and parameters: {'lambda_l1': 2.4032192398197603e-06, 'lambda_l2': 2.868804045681902e-06}. Best is trial 53 with value: 0.6878571684026484. regularization_factors, val_score: 0.687857: 75%|#######5 | 15/20 [00:37<00:12, 2.56s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000974 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577 [200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588 [300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856 [400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313 [500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084 [600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043 [700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014 Early stopping, best iteration is: [671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857 regularization_factors, val_score: 0.687857: 80%|######## | 16/20 [00:40<00:10, 2.56s/it][I 2020-09-27 05:00:06,891] Trial 58 finished with value: 0.6878571684147985 and parameters: {'lambda_l1': 3.739599166500671e-06, 'lambda_l2': 3.8681688367300176e-06}. Best is trial 53 with value: 0.6878571684026484. regularization_factors, val_score: 0.687857: 80%|######## | 16/20 [00:40<00:10, 2.56s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000993 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577 [200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588 [300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856 [400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313 [500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084 [600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043 [700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014 Early stopping, best iteration is: [671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857 regularization_factors, val_score: 0.687857: 85%|########5 | 17/20 [00:42<00:07, 2.56s/it][I 2020-09-27 05:00:09,442] Trial 59 finished with value: 0.6878571672506346 and parameters: {'lambda_l1': 2.703708966445098e-06, 'lambda_l2': 0.00019305898786142614}. Best is trial 59 with value: 0.6878571672506346. regularization_factors, val_score: 0.687857: 85%|########5 | 17/20 [00:42<00:07, 2.56s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000960 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577 [200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588 [300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856 [400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313 [500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084 [600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043 [700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014 Early stopping, best iteration is: [671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857 regularization_factors, val_score: 0.687857: 90%|######### | 18/20 [00:45<00:05, 2.56s/it][I 2020-09-27 05:00:11,996] Trial 60 finished with value: 0.6878571657826678 and parameters: {'lambda_l1': 9.271384341210084e-07, 'lambda_l2': 0.0004314431407377575}. Best is trial 60 with value: 0.6878571657826678. regularization_factors, val_score: 0.687857: 90%|######### | 18/20 [00:45<00:05, 2.56s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000998 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577 [200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588 [300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856 [400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313 [500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084 [600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043 [700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014 Early stopping, best iteration is: [671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857 regularization_factors, val_score: 0.687857: 95%|#########5| 19/20 [00:48<00:02, 2.56s/it][I 2020-09-27 05:00:14,562] Trial 61 finished with value: 0.6878571600694793 and parameters: {'lambda_l1': 8.018982307617414e-07, 'lambda_l2': 0.0013639030286262217}. Best is trial 61 with value: 0.6878571600694793. regularization_factors, val_score: 0.687857: 95%|#########5| 19/20 [00:48<00:02, 2.56s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000973 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577 [200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588 [300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856 [400] train's binary_logloss: 0.687189 valid's binary_logloss: 0.688313 [500] train's binary_logloss: 0.686756 valid's binary_logloss: 0.688084 [600] train's binary_logloss: 0.686406 valid's binary_logloss: 0.688043 [700] train's binary_logloss: 0.68613 valid's binary_logloss: 0.688014 Early stopping, best iteration is: [671] train's binary_logloss: 0.686204 valid's binary_logloss: 0.687857 regularization_factors, val_score: 0.687857: 100%|##########| 20/20 [00:50<00:00, 2.55s/it][I 2020-09-27 05:00:17,099] Trial 62 finished with value: 0.6878571590590913 and parameters: {'lambda_l1': 5.337492288420711e-07, 'lambda_l2': 0.0015292766949980237}. Best is trial 62 with value: 0.6878571590590913. regularization_factors, val_score: 0.687857: 100%|##########| 20/20 [00:50<00:00, 2.53s/it] min_data_in_leaf, val_score: 0.687857: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001704 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577 [200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588 [300] train's binary_logloss: 0.687741 valid's binary_logloss: 0.688856 [400] train's binary_logloss: 0.687194 valid's binary_logloss: 0.688269 [500] train's binary_logloss: 0.686757 valid's binary_logloss: 0.688121 [600] train's binary_logloss: 0.686414 valid's binary_logloss: 0.688068 Early stopping, best iteration is: [551] train's binary_logloss: 0.686573 valid's binary_logloss: 0.687983 min_data_in_leaf, val_score: 0.687857: 20%|## | 1/5 [00:02<00:08, 2.20s/it][I 2020-09-27 05:00:19,310] Trial 63 finished with value: 0.6879830688856906 and parameters: {'min_child_samples': 25}. Best is trial 63 with value: 0.6879830688856906. min_data_in_leaf, val_score: 0.687857: 20%|## | 1/5 [00:02<00:08, 2.20s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000979 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577 [200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588 [300] train's binary_logloss: 0.687742 valid's binary_logloss: 0.688876 [400] train's binary_logloss: 0.687187 valid's binary_logloss: 0.688301 [500] train's binary_logloss: 0.686743 valid's binary_logloss: 0.688115 [600] train's binary_logloss: 0.686394 valid's binary_logloss: 0.687995 Early stopping, best iteration is: [555] train's binary_logloss: 0.686546 valid's binary_logloss: 0.68795 min_data_in_leaf, val_score: 0.687857: 40%|#### | 2/5 [00:06<00:08, 2.72s/it][I 2020-09-27 05:00:23,240] Trial 64 finished with value: 0.687949637690665 and parameters: {'min_child_samples': 10}. Best is trial 64 with value: 0.687949637690665. min_data_in_leaf, val_score: 0.687857: 40%|#### | 2/5 [00:06<00:08, 2.72s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000925 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577 [200] train's binary_logloss: 0.688554 valid's binary_logloss: 0.689588 [300] train's binary_logloss: 0.687742 valid's binary_logloss: 0.688876 [400] train's binary_logloss: 0.687187 valid's binary_logloss: 0.688301 [500] train's binary_logloss: 0.686741 valid's binary_logloss: 0.688135 [600] train's binary_logloss: 0.686393 valid's binary_logloss: 0.688019 Early stopping, best iteration is: [553] train's binary_logloss: 0.68655 valid's binary_logloss: 0.687952 min_data_in_leaf, val_score: 0.687857: 60%|###### | 3/5 [00:08<00:05, 2.63s/it][I 2020-09-27 05:00:25,654] Trial 65 finished with value: 0.687952199445917 and parameters: {'min_child_samples': 5}. Best is trial 64 with value: 0.687949637690665. min_data_in_leaf, val_score: 0.687857: 60%|###### | 3/5 [00:08<00:05, 2.63s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001028 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68983 valid's binary_logloss: 0.690577 [200] train's binary_logloss: 0.688565 valid's binary_logloss: 0.689625 [300] train's binary_logloss: 0.687768 valid's binary_logloss: 0.688833 [400] train's binary_logloss: 0.687225 valid's binary_logloss: 0.688296 [500] train's binary_logloss: 0.686789 valid's binary_logloss: 0.688054 [600] train's binary_logloss: 0.686452 valid's binary_logloss: 0.687991 Early stopping, best iteration is: [555] train's binary_logloss: 0.6866 valid's binary_logloss: 0.687924 min_data_in_leaf, val_score: 0.687857: 80%|######## | 4/5 [00:10<00:02, 2.52s/it][I 2020-09-27 05:00:27,934] Trial 66 finished with value: 0.6879239560726775 and parameters: {'min_child_samples': 50}. Best is trial 66 with value: 0.6879239560726775. min_data_in_leaf, val_score: 0.687857: 80%|######## | 4/5 [00:10<00:02, 2.52s/it][LightGBM] [Info] Number of positive: 46279, number of negative: 46747 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000999 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497485 -> initscore=-0.010062 [LightGBM] [Info] Start training from score -0.010062 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689877 valid's binary_logloss: 0.690659 [200] train's binary_logloss: 0.688626 valid's binary_logloss: 0.689631 [300] train's binary_logloss: 0.687853 valid's binary_logloss: 0.68889 [400] train's binary_logloss: 0.687343 valid's binary_logloss: 0.688334 [500] train's binary_logloss: 0.686947 valid's binary_logloss: 0.688179 [600] train's binary_logloss: 0.686609 valid's binary_logloss: 0.688099 Early stopping, best iteration is: [555] train's binary_logloss: 0.686758 valid's binary_logloss: 0.688011 min_data_in_leaf, val_score: 0.687857: 100%|##########| 5/5 [00:13<00:00, 2.42s/it][I 2020-09-27 05:00:30,120] Trial 67 finished with value: 0.6880111488280128 and parameters: {'min_child_samples': 100}. Best is trial 66 with value: 0.6879239560726775. min_data_in_leaf, val_score: 0.687857: 100%|##########| 5/5 [00:13<00:00, 2.60s/it] Fold : 7 [I 2020-09-27 05:00:30,278] A new study created in memory with name: no-name-ad8f4155-35d8-4bc1-80b9-1a44d36999f7 feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000941 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665204 valid's binary_logloss: 0.691098 Early stopping, best iteration is: [38] train's binary_logloss: 0.679318 valid's binary_logloss: 0.689859 feature_fraction, val_score: 0.689859: 14%|#4 | 1/7 [00:00<00:05, 1.15it/s][I 2020-09-27 05:00:31,155] Trial 0 finished with value: 0.6898590878905122 and parameters: {'feature_fraction': 0.5}. Best is trial 0 with value: 0.6898590878905122. feature_fraction, val_score: 0.689859: 14%|#4 | 1/7 [00:00<00:05, 1.15it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000919 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.66602 valid's binary_logloss: 0.689895 Early stopping, best iteration is: [60] train's binary_logloss: 0.674409 valid's binary_logloss: 0.689601 feature_fraction, val_score: 0.689601: 29%|##8 | 2/7 [00:01<00:04, 1.20it/s][I 2020-09-27 05:00:31,908] Trial 1 finished with value: 0.6896011115289385 and parameters: {'feature_fraction': 0.4}. Best is trial 1 with value: 0.6896011115289385. feature_fraction, val_score: 0.689601: 29%|##8 | 2/7 [00:01<00:04, 1.20it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001642 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.662693 valid's binary_logloss: 0.690012 Early stopping, best iteration is: [39] train's binary_logloss: 0.677876 valid's binary_logloss: 0.689256 feature_fraction, val_score: 0.689256: 43%|####2 | 3/7 [00:02<00:03, 1.23it/s][I 2020-09-27 05:00:32,673] Trial 2 finished with value: 0.6892558819835889 and parameters: {'feature_fraction': 1.0}. Best is trial 2 with value: 0.6892558819835889. feature_fraction, val_score: 0.689256: 43%|####2 | 3/7 [00:02<00:03, 1.23it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001641 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663231 valid's binary_logloss: 0.690116 Early stopping, best iteration is: [31] train's binary_logloss: 0.6803 valid's binary_logloss: 0.689625 feature_fraction, val_score: 0.689256: 57%|#####7 | 4/7 [00:03<00:02, 1.26it/s][I 2020-09-27 05:00:33,421] Trial 3 finished with value: 0.6896254736016139 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 2 with value: 0.6892558819835889. feature_fraction, val_score: 0.689256: 57%|#####7 | 4/7 [00:03<00:02, 1.26it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012212 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664409 valid's binary_logloss: 0.68913 [200] train's binary_logloss: 0.644514 valid's binary_logloss: 0.690437 Early stopping, best iteration is: [100] train's binary_logloss: 0.664409 valid's binary_logloss: 0.68913 feature_fraction, val_score: 0.689130: 71%|#######1 | 5/7 [00:04<00:01, 1.20it/s][I 2020-09-27 05:00:34,351] Trial 4 finished with value: 0.6891302495722187 and parameters: {'feature_fraction': 0.6}. Best is trial 4 with value: 0.6891302495722187. feature_fraction, val_score: 0.689130: 71%|#######1 | 5/7 [00:04<00:01, 1.20it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016528 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664259 valid's binary_logloss: 0.689566 Early stopping, best iteration is: [58] train's binary_logloss: 0.673563 valid's binary_logloss: 0.688962 feature_fraction, val_score: 0.688962: 86%|########5 | 6/7 [00:04<00:00, 1.21it/s][I 2020-09-27 05:00:35,160] Trial 5 finished with value: 0.6889623887000853 and parameters: {'feature_fraction': 0.7}. Best is trial 5 with value: 0.6889623887000853. feature_fraction, val_score: 0.688962: 86%|########5 | 6/7 [00:04<00:00, 1.21it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001610 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663479 valid's binary_logloss: 0.689779 Early stopping, best iteration is: [60] train's binary_logloss: 0.67284 valid's binary_logloss: 0.689141 feature_fraction, val_score: 0.688962: 100%|##########| 7/7 [00:05<00:00, 1.19it/s][I 2020-09-27 05:00:36,033] Trial 6 finished with value: 0.6891408826755168 and parameters: {'feature_fraction': 0.8}. Best is trial 5 with value: 0.6889623887000853. feature_fraction, val_score: 0.688962: 100%|##########| 7/7 [00:05<00:00, 1.22it/s] num_leaves, val_score: 0.688962: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002027 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.552919 valid's binary_logloss: 0.697247 Early stopping, best iteration is: [23] train's binary_logloss: 0.648261 valid's binary_logloss: 0.691036 num_leaves, val_score: 0.688962: 5%|5 | 1/20 [00:01<00:27, 1.44s/it][I 2020-09-27 05:00:37,483] Trial 7 finished with value: 0.6910358844305945 and parameters: {'num_leaves': 211}. Best is trial 7 with value: 0.6910358844305945. num_leaves, val_score: 0.688962: 5%|5 | 1/20 [00:01<00:27, 1.44s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007979 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.62836 valid's binary_logloss: 0.691669 Early stopping, best iteration is: [27] train's binary_logloss: 0.669575 valid's binary_logloss: 0.689768 num_leaves, val_score: 0.688962: 10%|# | 2/20 [00:02<00:22, 1.25s/it][I 2020-09-27 05:00:38,292] Trial 8 finished with value: 0.6897684410879633 and parameters: {'num_leaves': 81}. Best is trial 8 with value: 0.6897684410879633. num_leaves, val_score: 0.688962: 10%|# | 2/20 [00:02<00:22, 1.25s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001953 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.550486 valid's binary_logloss: 0.694876 Early stopping, best iteration is: [27] train's binary_logloss: 0.641202 valid's binary_logloss: 0.689961 num_leaves, val_score: 0.688962: 15%|#5 | 3/20 [00:03<00:22, 1.31s/it][I 2020-09-27 05:00:39,748] Trial 9 finished with value: 0.6899614185647561 and parameters: {'num_leaves': 215}. Best is trial 8 with value: 0.6897684410879633. num_leaves, val_score: 0.688962: 15%|#5 | 3/20 [00:03<00:22, 1.31s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012913 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685219 valid's binary_logloss: 0.689356 [200] train's binary_logloss: 0.681241 valid's binary_logloss: 0.689034 [300] train's binary_logloss: 0.677713 valid's binary_logloss: 0.689095 Early stopping, best iteration is: [228] train's binary_logloss: 0.680209 valid's binary_logloss: 0.688922 num_leaves, val_score: 0.688922: 20%|## | 4/20 [00:04<00:20, 1.28s/it][I 2020-09-27 05:00:40,941] Trial 10 finished with value: 0.6889216574404491 and parameters: {'num_leaves': 6}. Best is trial 10 with value: 0.6889216574404491. num_leaves, val_score: 0.688922: 20%|## | 4/20 [00:04<00:20, 1.28s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.006570 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687266 valid's binary_logloss: 0.68936 [200] train's binary_logloss: 0.684753 valid's binary_logloss: 0.688801 [300] train's binary_logloss: 0.682767 valid's binary_logloss: 0.688695 Early stopping, best iteration is: [259] train's binary_logloss: 0.683544 valid's binary_logloss: 0.688542 num_leaves, val_score: 0.688542: 25%|##5 | 5/20 [00:06<00:19, 1.29s/it][I 2020-09-27 05:00:42,272] Trial 11 finished with value: 0.6885421835424413 and parameters: {'num_leaves': 4}. Best is trial 11 with value: 0.6885421835424413. num_leaves, val_score: 0.688542: 25%|##5 | 5/20 [00:06<00:19, 1.29s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012743 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.683209 valid's binary_logloss: 0.688991 [200] train's binary_logloss: 0.677714 valid's binary_logloss: 0.689043 Early stopping, best iteration is: [167] train's binary_logloss: 0.679404 valid's binary_logloss: 0.688853 num_leaves, val_score: 0.688542: 30%|### | 6/20 [00:07<00:17, 1.22s/it][I 2020-09-27 05:00:43,311] Trial 12 finished with value: 0.6888528740217242 and parameters: {'num_leaves': 8}. Best is trial 11 with value: 0.6885421835424413. num_leaves, val_score: 0.688542: 30%|### | 6/20 [00:07<00:17, 1.22s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.022273 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687266 valid's binary_logloss: 0.68936 [200] train's binary_logloss: 0.684753 valid's binary_logloss: 0.688801 [300] train's binary_logloss: 0.682767 valid's binary_logloss: 0.688695 Early stopping, best iteration is: [259] train's binary_logloss: 0.683544 valid's binary_logloss: 0.688542 num_leaves, val_score: 0.688542: 35%|###5 | 7/20 [00:08<00:16, 1.24s/it][I 2020-09-27 05:00:44,603] Trial 13 finished with value: 0.6885421835424413 and parameters: {'num_leaves': 4}. Best is trial 11 with value: 0.6885421835424413. num_leaves, val_score: 0.688542: 35%|###5 | 7/20 [00:08<00:16, 1.24s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012970 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.636905 valid's binary_logloss: 0.691529 Early stopping, best iteration is: [47] train's binary_logloss: 0.661435 valid's binary_logloss: 0.689547 num_leaves, val_score: 0.688542: 40%|#### | 8/20 [00:09<00:13, 1.14s/it][I 2020-09-27 05:00:45,518] Trial 14 finished with value: 0.6895472826915808 and parameters: {'num_leaves': 69}. Best is trial 11 with value: 0.6885421835424413. num_leaves, val_score: 0.688542: 40%|#### | 8/20 [00:09<00:13, 1.14s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008973 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68623 valid's binary_logloss: 0.689472 [200] train's binary_logloss: 0.682968 valid's binary_logloss: 0.689396 Early stopping, best iteration is: [166] train's binary_logloss: 0.683998 valid's binary_logloss: 0.689301 num_leaves, val_score: 0.688542: 45%|####5 | 9/20 [00:10<00:12, 1.09s/it][I 2020-09-27 05:00:46,499] Trial 15 finished with value: 0.6893008615554372 and parameters: {'num_leaves': 5}. Best is trial 11 with value: 0.6885421835424413. num_leaves, val_score: 0.688542: 45%|####5 | 9/20 [00:10<00:12, 1.09s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008097 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.641155 valid's binary_logloss: 0.691233 Early stopping, best iteration is: [44] train's binary_logloss: 0.665246 valid's binary_logloss: 0.689534 num_leaves, val_score: 0.688542: 50%|##### | 10/20 [00:11<00:10, 1.03s/it][I 2020-09-27 05:00:47,384] Trial 16 finished with value: 0.689534420595521 and parameters: {'num_leaves': 63}. Best is trial 11 with value: 0.6885421835424413. num_leaves, val_score: 0.688542: 50%|##### | 10/20 [00:11<00:10, 1.03s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007854 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.594183 valid's binary_logloss: 0.693385 Early stopping, best iteration is: [30] train's binary_logloss: 0.653722 valid's binary_logloss: 0.690977 num_leaves, val_score: 0.688542: 55%|#####5 | 11/20 [00:12<00:09, 1.04s/it][I 2020-09-27 05:00:48,441] Trial 17 finished with value: 0.6909774333414871 and parameters: {'num_leaves': 139}. Best is trial 11 with value: 0.6885421835424413. num_leaves, val_score: 0.688542: 55%|#####5 | 11/20 [00:12<00:09, 1.04s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.014042 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.660928 valid's binary_logloss: 0.689724 Early stopping, best iteration is: [65] train's binary_logloss: 0.669654 valid's binary_logloss: 0.689001 num_leaves, val_score: 0.688542: 60%|###### | 12/20 [00:13<00:07, 1.02it/s][I 2020-09-27 05:00:49,287] Trial 18 finished with value: 0.6890011375101475 and parameters: {'num_leaves': 35}. Best is trial 11 with value: 0.6885421835424413. num_leaves, val_score: 0.688542: 60%|###### | 12/20 [00:13<00:07, 1.02it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007812 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.594516 valid's binary_logloss: 0.692963 Early stopping, best iteration is: [16] train's binary_logloss: 0.670225 valid's binary_logloss: 0.691015 num_leaves, val_score: 0.688542: 65%|######5 | 13/20 [00:14<00:06, 1.03it/s][I 2020-09-27 05:00:50,248] Trial 19 finished with value: 0.6910149125364163 and parameters: {'num_leaves': 136}. Best is trial 11 with value: 0.6885421835424413. num_leaves, val_score: 0.688542: 65%|######5 | 13/20 [00:14<00:06, 1.03it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008513 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.614802 valid's binary_logloss: 0.691487 Early stopping, best iteration is: [50] train's binary_logloss: 0.64677 valid's binary_logloss: 0.690169 num_leaves, val_score: 0.688542: 70%|####### | 14/20 [00:15<00:05, 1.00it/s][I 2020-09-27 05:00:51,302] Trial 20 finished with value: 0.6901693018069993 and parameters: {'num_leaves': 103}. Best is trial 11 with value: 0.6885421835424413. num_leaves, val_score: 0.688542: 70%|####### | 14/20 [00:15<00:05, 1.00it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001631 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68623 valid's binary_logloss: 0.689472 [200] train's binary_logloss: 0.682968 valid's binary_logloss: 0.689396 Early stopping, best iteration is: [166] train's binary_logloss: 0.683998 valid's binary_logloss: 0.689301 num_leaves, val_score: 0.688542: 75%|#######5 | 15/20 [00:16<00:05, 1.00s/it][I 2020-09-27 05:00:52,310] Trial 21 finished with value: 0.6893008615554372 and parameters: {'num_leaves': 5}. Best is trial 11 with value: 0.6885421835424413. num_leaves, val_score: 0.688542: 75%|#######5 | 15/20 [00:16<00:05, 1.00s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001749 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663209 valid's binary_logloss: 0.689372 Early stopping, best iteration is: [56] train's binary_logloss: 0.673672 valid's binary_logloss: 0.688635 num_leaves, val_score: 0.688542: 80%|######## | 16/20 [00:17<00:03, 1.04it/s][I 2020-09-27 05:00:53,172] Trial 22 finished with value: 0.6886345495846673 and parameters: {'num_leaves': 32}. Best is trial 11 with value: 0.6885421835424413. num_leaves, val_score: 0.688542: 80%|######## | 16/20 [00:17<00:03, 1.04it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008554 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.65661 valid's binary_logloss: 0.689805 Early stopping, best iteration is: [44] train's binary_logloss: 0.673206 valid's binary_logloss: 0.689461 num_leaves, val_score: 0.688542: 85%|########5 | 17/20 [00:17<00:02, 1.10it/s][I 2020-09-27 05:00:53,968] Trial 23 finished with value: 0.6894608609879804 and parameters: {'num_leaves': 41}. Best is trial 11 with value: 0.6885421835424413. num_leaves, val_score: 0.688542: 85%|########5 | 17/20 [00:17<00:02, 1.10it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012405 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.658212 valid's binary_logloss: 0.690617 Early stopping, best iteration is: [33] train's binary_logloss: 0.677455 valid's binary_logloss: 0.689576 num_leaves, val_score: 0.688542: 90%|######### | 18/20 [00:18<00:01, 1.16it/s][I 2020-09-27 05:00:54,710] Trial 24 finished with value: 0.6895760118681855 and parameters: {'num_leaves': 39}. Best is trial 11 with value: 0.6885421835424413. num_leaves, val_score: 0.688542: 90%|######### | 18/20 [00:18<00:01, 1.16it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008295 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.577324 valid's binary_logloss: 0.696256 Early stopping, best iteration is: [31] train's binary_logloss: 0.645438 valid's binary_logloss: 0.691006 num_leaves, val_score: 0.688542: 95%|#########5| 19/20 [00:19<00:00, 1.03it/s][I 2020-09-27 05:00:55,935] Trial 25 finished with value: 0.6910055064236426 and parameters: {'num_leaves': 167}. Best is trial 11 with value: 0.6885421835424413. num_leaves, val_score: 0.688542: 95%|#########5| 19/20 [00:19<00:00, 1.03it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012736 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667178 valid's binary_logloss: 0.689846 Early stopping, best iteration is: [48] train's binary_logloss: 0.677808 valid's binary_logloss: 0.689292 num_leaves, val_score: 0.688542: 100%|##########| 20/20 [00:20<00:00, 1.11it/s][I 2020-09-27 05:00:56,664] Trial 26 finished with value: 0.6892917155718229 and parameters: {'num_leaves': 27}. Best is trial 11 with value: 0.6885421835424413. num_leaves, val_score: 0.688542: 100%|##########| 20/20 [00:20<00:00, 1.03s/it] bagging, val_score: 0.688542: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008116 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687216 valid's binary_logloss: 0.689193 [200] train's binary_logloss: 0.684606 valid's binary_logloss: 0.688599 [300] train's binary_logloss: 0.682514 valid's binary_logloss: 0.688495 Early stopping, best iteration is: [243] train's binary_logloss: 0.683636 valid's binary_logloss: 0.688314 bagging, val_score: 0.688314: 10%|# | 1/10 [00:01<00:11, 1.33s/it][I 2020-09-27 05:00:58,010] Trial 27 finished with value: 0.6883135095365117 and parameters: {'bagging_fraction': 0.8631686996360303, 'bagging_freq': 3}. Best is trial 27 with value: 0.6883135095365117. bagging, val_score: 0.688314: 10%|# | 1/10 [00:01<00:11, 1.33s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008381 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687216 valid's binary_logloss: 0.689219 [200] train's binary_logloss: 0.684738 valid's binary_logloss: 0.688577 [300] train's binary_logloss: 0.682586 valid's binary_logloss: 0.688474 [400] train's binary_logloss: 0.680625 valid's binary_logloss: 0.688466 Early stopping, best iteration is: [388] train's binary_logloss: 0.680824 valid's binary_logloss: 0.688307 bagging, val_score: 0.688307: 20%|## | 2/10 [00:03<00:12, 1.50s/it][I 2020-09-27 05:00:59,908] Trial 28 finished with value: 0.6883067782443618 and parameters: {'bagging_fraction': 0.8747424021051551, 'bagging_freq': 3}. Best is trial 28 with value: 0.6883067782443618. bagging, val_score: 0.688307: 20%|## | 2/10 [00:03<00:12, 1.50s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008348 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687228 valid's binary_logloss: 0.689193 [200] train's binary_logloss: 0.684624 valid's binary_logloss: 0.688721 [300] train's binary_logloss: 0.682522 valid's binary_logloss: 0.688655 [400] train's binary_logloss: 0.680572 valid's binary_logloss: 0.688787 Early stopping, best iteration is: [313] train's binary_logloss: 0.682272 valid's binary_logloss: 0.688627 bagging, val_score: 0.688307: 30%|### | 3/10 [00:04<00:10, 1.54s/it][I 2020-09-27 05:01:01,550] Trial 29 finished with value: 0.6886270116658745 and parameters: {'bagging_fraction': 0.8767026255324745, 'bagging_freq': 3}. Best is trial 28 with value: 0.6883067782443618. bagging, val_score: 0.688307: 30%|### | 3/10 [00:04<00:10, 1.54s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008608 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687185 valid's binary_logloss: 0.689143 [200] train's binary_logloss: 0.684623 valid's binary_logloss: 0.688512 [300] train's binary_logloss: 0.682521 valid's binary_logloss: 0.688305 Early stopping, best iteration is: [299] train's binary_logloss: 0.682543 valid's binary_logloss: 0.688237 bagging, val_score: 0.688237: 40%|#### | 4/10 [00:06<00:09, 1.55s/it][I 2020-09-27 05:01:03,118] Trial 30 finished with value: 0.6882368432610401 and parameters: {'bagging_fraction': 0.8508850031981232, 'bagging_freq': 3}. Best is trial 30 with value: 0.6882368432610401. bagging, val_score: 0.688237: 40%|#### | 4/10 [00:06<00:09, 1.55s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013843 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687182 valid's binary_logloss: 0.689138 [200] train's binary_logloss: 0.684621 valid's binary_logloss: 0.688476 [300] train's binary_logloss: 0.682518 valid's binary_logloss: 0.688405 [400] train's binary_logloss: 0.680515 valid's binary_logloss: 0.688508 Early stopping, best iteration is: [307] train's binary_logloss: 0.682384 valid's binary_logloss: 0.688333 bagging, val_score: 0.688237: 50%|##### | 5/10 [00:08<00:07, 1.57s/it][I 2020-09-27 05:01:04,736] Trial 31 finished with value: 0.6883329240332864 and parameters: {'bagging_fraction': 0.8593267514380051, 'bagging_freq': 3}. Best is trial 30 with value: 0.6882368432610401. bagging, val_score: 0.688237: 50%|##### | 5/10 [00:08<00:07, 1.57s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010252 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687215 valid's binary_logloss: 0.689327 [200] train's binary_logloss: 0.684669 valid's binary_logloss: 0.688643 [300] train's binary_logloss: 0.682522 valid's binary_logloss: 0.68885 [400] train's binary_logloss: 0.680567 valid's binary_logloss: 0.688669 Early stopping, best iteration is: [353] train's binary_logloss: 0.681477 valid's binary_logloss: 0.68855 bagging, val_score: 0.688237: 60%|###### | 6/10 [00:09<00:06, 1.62s/it][I 2020-09-27 05:01:06,469] Trial 32 finished with value: 0.6885496852605667 and parameters: {'bagging_fraction': 0.8612598431625776, 'bagging_freq': 3}. Best is trial 30 with value: 0.6882368432610401. bagging, val_score: 0.688237: 60%|###### | 6/10 [00:09<00:06, 1.62s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008437 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68717 valid's binary_logloss: 0.689286 [200] train's binary_logloss: 0.684607 valid's binary_logloss: 0.688529 [300] train's binary_logloss: 0.682498 valid's binary_logloss: 0.688434 Early stopping, best iteration is: [262] train's binary_logloss: 0.683269 valid's binary_logloss: 0.688293 bagging, val_score: 0.688237: 70%|####### | 7/10 [00:11<00:04, 1.57s/it][I 2020-09-27 05:01:07,911] Trial 33 finished with value: 0.6882931441850038 and parameters: {'bagging_fraction': 0.8671876569591457, 'bagging_freq': 3}. Best is trial 30 with value: 0.6882368432610401. bagging, val_score: 0.688237: 70%|####### | 7/10 [00:11<00:04, 1.57s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008157 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687191 valid's binary_logloss: 0.68922 [200] train's binary_logloss: 0.68462 valid's binary_logloss: 0.688602 Early stopping, best iteration is: [186] train's binary_logloss: 0.684928 valid's binary_logloss: 0.688549 bagging, val_score: 0.688237: 80%|######## | 8/10 [00:12<00:02, 1.45s/it][I 2020-09-27 05:01:09,099] Trial 34 finished with value: 0.6885490341442905 and parameters: {'bagging_fraction': 0.8804920945207426, 'bagging_freq': 3}. Best is trial 30 with value: 0.6882368432610401. bagging, val_score: 0.688237: 80%|######## | 8/10 [00:12<00:02, 1.45s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013640 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68703 valid's binary_logloss: 0.688776 [200] train's binary_logloss: 0.68453 valid's binary_logloss: 0.688164 [300] train's binary_logloss: 0.682427 valid's binary_logloss: 0.687911 [400] train's binary_logloss: 0.680592 valid's binary_logloss: 0.688176 Early stopping, best iteration is: [347] train's binary_logloss: 0.681525 valid's binary_logloss: 0.687729 bagging, val_score: 0.687729: 90%|######### | 9/10 [00:14<00:01, 1.52s/it][I 2020-09-27 05:01:10,771] Trial 35 finished with value: 0.6877285343949435 and parameters: {'bagging_fraction': 0.6813927760377331, 'bagging_freq': 3}. Best is trial 35 with value: 0.6877285343949435. bagging, val_score: 0.687729: 90%|######### | 9/10 [00:14<00:01, 1.52s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008939 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687118 valid's binary_logloss: 0.688984 [200] train's binary_logloss: 0.684488 valid's binary_logloss: 0.688477 [300] train's binary_logloss: 0.682515 valid's binary_logloss: 0.688435 [400] train's binary_logloss: 0.680676 valid's binary_logloss: 0.688358 [500] train's binary_logloss: 0.678963 valid's binary_logloss: 0.688512 Early stopping, best iteration is: [456] train's binary_logloss: 0.679669 valid's binary_logloss: 0.688231 bagging, val_score: 0.687729: 100%|##########| 10/10 [00:15<00:00, 1.63s/it][I 2020-09-27 05:01:12,671] Trial 36 finished with value: 0.6882307476214391 and parameters: {'bagging_fraction': 0.6088933564022119, 'bagging_freq': 6}. Best is trial 35 with value: 0.6877285343949435. bagging, val_score: 0.687729: 100%|##########| 10/10 [00:16<00:00, 1.60s/it] feature_fraction_stage2, val_score: 0.687729: 0%| | 0/6 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001488 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687028 valid's binary_logloss: 0.688811 [200] train's binary_logloss: 0.684577 valid's binary_logloss: 0.68832 [300] train's binary_logloss: 0.682495 valid's binary_logloss: 0.688292 Early stopping, best iteration is: [229] train's binary_logloss: 0.683926 valid's binary_logloss: 0.688061 feature_fraction_stage2, val_score: 0.687729: 17%|#6 | 1/6 [00:01<00:06, 1.31s/it][I 2020-09-27 05:01:13,995] Trial 37 finished with value: 0.6880605102479459 and parameters: {'feature_fraction': 0.7799999999999999}. Best is trial 37 with value: 0.6880605102479459. feature_fraction_stage2, val_score: 0.687729: 17%|#6 | 1/6 [00:01<00:06, 1.31s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007740 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68703 valid's binary_logloss: 0.688776 [200] train's binary_logloss: 0.68453 valid's binary_logloss: 0.688164 [300] train's binary_logloss: 0.682427 valid's binary_logloss: 0.687911 [400] train's binary_logloss: 0.680592 valid's binary_logloss: 0.688176 Early stopping, best iteration is: [347] train's binary_logloss: 0.681525 valid's binary_logloss: 0.687729 feature_fraction_stage2, val_score: 0.687729: 33%|###3 | 2/6 [00:03<00:05, 1.43s/it][I 2020-09-27 05:01:15,700] Trial 38 finished with value: 0.6877285343949435 and parameters: {'feature_fraction': 0.6839999999999999}. Best is trial 38 with value: 0.6877285343949435. feature_fraction_stage2, val_score: 0.687729: 33%|###3 | 2/6 [00:03<00:05, 1.43s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009068 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68704 valid's binary_logloss: 0.688886 [200] train's binary_logloss: 0.684508 valid's binary_logloss: 0.688072 [300] train's binary_logloss: 0.682367 valid's binary_logloss: 0.688022 Early stopping, best iteration is: [230] train's binary_logloss: 0.683826 valid's binary_logloss: 0.687901 feature_fraction_stage2, val_score: 0.687729: 50%|##### | 3/6 [00:04<00:04, 1.41s/it][I 2020-09-27 05:01:17,068] Trial 39 finished with value: 0.68790080420035 and parameters: {'feature_fraction': 0.716}. Best is trial 38 with value: 0.6877285343949435. feature_fraction_stage2, val_score: 0.687729: 50%|##### | 3/6 [00:04<00:04, 1.41s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008155 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68704 valid's binary_logloss: 0.688886 [200] train's binary_logloss: 0.684508 valid's binary_logloss: 0.688072 [300] train's binary_logloss: 0.682367 valid's binary_logloss: 0.688022 Early stopping, best iteration is: [230] train's binary_logloss: 0.683826 valid's binary_logloss: 0.687901 feature_fraction_stage2, val_score: 0.687729: 67%|######6 | 4/6 [00:05<00:02, 1.40s/it][I 2020-09-27 05:01:18,433] Trial 40 finished with value: 0.68790080420035 and parameters: {'feature_fraction': 0.748}. Best is trial 38 with value: 0.6877285343949435. feature_fraction_stage2, val_score: 0.687729: 67%|######6 | 4/6 [00:05<00:02, 1.40s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010780 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737 [200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967 [300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694 Early stopping, best iteration is: [231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553 feature_fraction_stage2, val_score: 0.687553: 83%|########3 | 5/6 [00:06<00:01, 1.35s/it][I 2020-09-27 05:01:19,668] Trial 41 finished with value: 0.6875532234561516 and parameters: {'feature_fraction': 0.652}. Best is trial 41 with value: 0.6875532234561516. feature_fraction_stage2, val_score: 0.687553: 83%|########3 | 5/6 [00:06<00:01, 1.35s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.015763 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687015 valid's binary_logloss: 0.68889 [200] train's binary_logloss: 0.684543 valid's binary_logloss: 0.688419 [300] train's binary_logloss: 0.682467 valid's binary_logloss: 0.688343 Early stopping, best iteration is: [234] train's binary_logloss: 0.683801 valid's binary_logloss: 0.688055 feature_fraction_stage2, val_score: 0.687553: 100%|##########| 6/6 [00:08<00:00, 1.32s/it][I 2020-09-27 05:01:20,912] Trial 42 finished with value: 0.6880548224298009 and parameters: {'feature_fraction': 0.62}. Best is trial 41 with value: 0.6875532234561516. feature_fraction_stage2, val_score: 0.687553: 100%|##########| 6/6 [00:08<00:00, 1.37s/it] regularization_factors, val_score: 0.687553: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009634 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737 [200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967 [300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694 Early stopping, best iteration is: [231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553 regularization_factors, val_score: 0.687553: 5%|5 | 1/20 [00:01<00:24, 1.28s/it][I 2020-09-27 05:01:22,212] Trial 43 finished with value: 0.6875532234753943 and parameters: {'lambda_l1': 8.067143310213936e-07, 'lambda_l2': 1.7924277356165816e-07}. Best is trial 43 with value: 0.6875532234753943. regularization_factors, val_score: 0.687553: 5%|5 | 1/20 [00:01<00:24, 1.28s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.005038 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737 [200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967 [300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694 Early stopping, best iteration is: [231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553 regularization_factors, val_score: 0.687553: 10%|# | 2/20 [00:02<00:23, 1.30s/it][I 2020-09-27 05:01:23,564] Trial 44 finished with value: 0.6875532234639651 and parameters: {'lambda_l1': 2.9817465406512945e-07, 'lambda_l2': 7.314463336600674e-08}. Best is trial 44 with value: 0.6875532234639651. regularization_factors, val_score: 0.687553: 10%|# | 2/20 [00:02<00:23, 1.30s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012896 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737 [200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967 [300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694 Early stopping, best iteration is: [231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553 regularization_factors, val_score: 0.687553: 15%|#5 | 3/20 [00:03<00:21, 1.29s/it][I 2020-09-27 05:01:24,835] Trial 45 finished with value: 0.6875532234634629 and parameters: {'lambda_l1': 2.9860158446650325e-07, 'lambda_l2': 5.2722857836715734e-08}. Best is trial 45 with value: 0.6875532234634629. regularization_factors, val_score: 0.687553: 15%|#5 | 3/20 [00:03<00:21, 1.29s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007909 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737 [200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967 [300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694 Early stopping, best iteration is: [231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553 regularization_factors, val_score: 0.687553: 20%|## | 4/20 [00:05<00:20, 1.28s/it][I 2020-09-27 05:01:26,067] Trial 46 finished with value: 0.6875532234640681 and parameters: {'lambda_l1': 3.287625005492051e-07, 'lambda_l2': 3.053523237410299e-08}. Best is trial 45 with value: 0.6875532234634629. regularization_factors, val_score: 0.687553: 20%|## | 4/20 [00:05<00:20, 1.28s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007843 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737 [200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967 [300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694 Early stopping, best iteration is: [231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553 regularization_factors, val_score: 0.687553: 25%|##5 | 5/20 [00:06<00:19, 1.27s/it][I 2020-09-27 05:01:27,332] Trial 47 finished with value: 0.6875532234679212 and parameters: {'lambda_l1': 4.737421684219058e-07, 'lambda_l2': 3.2678236762325496e-08}. Best is trial 45 with value: 0.6875532234634629. regularization_factors, val_score: 0.687553: 25%|##5 | 5/20 [00:06<00:19, 1.27s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012460 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737 [200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967 [300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694 Early stopping, best iteration is: [231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553 regularization_factors, val_score: 0.687553: 30%|### | 6/20 [00:07<00:17, 1.27s/it][I 2020-09-27 05:01:28,588] Trial 48 finished with value: 0.6875532234603976 and parameters: {'lambda_l1': 1.8616761378629058e-07, 'lambda_l2': 1.614659705144355e-08}. Best is trial 48 with value: 0.6875532234603976. regularization_factors, val_score: 0.687553: 30%|### | 6/20 [00:07<00:17, 1.27s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008626 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737 [200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967 [300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694 Early stopping, best iteration is: [231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553 regularization_factors, val_score: 0.687553: 35%|###5 | 7/20 [00:08<00:16, 1.27s/it][I 2020-09-27 05:01:29,854] Trial 49 finished with value: 0.6875532234594542 and parameters: {'lambda_l1': 1.3735113285620687e-07, 'lambda_l2': 2.067788810166821e-08}. Best is trial 49 with value: 0.6875532234594542. regularization_factors, val_score: 0.687553: 35%|###5 | 7/20 [00:08<00:16, 1.27s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012441 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737 [200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967 [300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694 Early stopping, best iteration is: [231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553 regularization_factors, val_score: 0.687553: 40%|#### | 8/20 [00:10<00:15, 1.28s/it][I 2020-09-27 05:01:31,172] Trial 50 finished with value: 0.6875532234565727 and parameters: {'lambda_l1': 2.5154633280823624e-08, 'lambda_l2': 1.0923630752094391e-08}. Best is trial 50 with value: 0.6875532234565727. regularization_factors, val_score: 0.687553: 40%|#### | 8/20 [00:10<00:15, 1.28s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002245 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737 [200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967 [300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694 Early stopping, best iteration is: [231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553 regularization_factors, val_score: 0.687553: 45%|####5 | 9/20 [00:11<00:14, 1.29s/it][I 2020-09-27 05:01:32,467] Trial 51 finished with value: 0.6875532234566062 and parameters: {'lambda_l1': 3.0807637392676064e-08, 'lambda_l2': 1.4533537232191534e-08}. Best is trial 50 with value: 0.6875532234565727. regularization_factors, val_score: 0.687553: 45%|####5 | 9/20 [00:11<00:14, 1.29s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008797 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737 [200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967 [300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694 Early stopping, best iteration is: [231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553 regularization_factors, val_score: 0.687553: 50%|##### | 10/20 [00:12<00:12, 1.28s/it][I 2020-09-27 05:01:33,731] Trial 52 finished with value: 0.687553223456156 and parameters: {'lambda_l1': 1.0265554714929266e-08, 'lambda_l2': 1.0447547951042816e-08}. Best is trial 52 with value: 0.687553223456156. regularization_factors, val_score: 0.687553: 50%|##### | 10/20 [00:12<00:12, 1.28s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001950 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737 [200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967 [300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694 Early stopping, best iteration is: [231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553 regularization_factors, val_score: 0.687553: 55%|#####5 | 11/20 [00:14<00:11, 1.30s/it][I 2020-09-27 05:01:35,074] Trial 53 finished with value: 0.6875532234561593 and parameters: {'lambda_l1': 1.0392524686592004e-08, 'lambda_l2': 1.0663353352190143e-08}. Best is trial 52 with value: 0.687553223456156. regularization_factors, val_score: 0.687553: 55%|#####5 | 11/20 [00:14<00:11, 1.30s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008108 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687064 valid's binary_logloss: 0.688739 [200] train's binary_logloss: 0.684486 valid's binary_logloss: 0.688006 [300] train's binary_logloss: 0.682501 valid's binary_logloss: 0.687947 Early stopping, best iteration is: [231] train's binary_logloss: 0.683839 valid's binary_logloss: 0.687771 regularization_factors, val_score: 0.687553: 60%|###### | 12/20 [00:15<00:10, 1.29s/it][I 2020-09-27 05:01:36,332] Trial 54 finished with value: 0.6877705631041466 and parameters: {'lambda_l1': 2.1247372830919867e-08, 'lambda_l2': 0.06309677087509566}. Best is trial 52 with value: 0.687553223456156. regularization_factors, val_score: 0.687553: 60%|###### | 12/20 [00:15<00:10, 1.29s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011180 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737 [200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967 [300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694 Early stopping, best iteration is: [231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553 regularization_factors, val_score: 0.687553: 65%|######5 | 13/20 [00:16<00:08, 1.28s/it][I 2020-09-27 05:01:37,586] Trial 55 finished with value: 0.6875532234562124 and parameters: {'lambda_l1': 1.0739708350048745e-08, 'lambda_l2': 1.1070314813782383e-08}. Best is trial 52 with value: 0.687553223456156. regularization_factors, val_score: 0.687553: 65%|######5 | 13/20 [00:16<00:08, 1.28s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008356 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687053 valid's binary_logloss: 0.68877 [200] train's binary_logloss: 0.684524 valid's binary_logloss: 0.688207 [300] train's binary_logloss: 0.682531 valid's binary_logloss: 0.687938 [400] train's binary_logloss: 0.68063 valid's binary_logloss: 0.688065 Early stopping, best iteration is: [348] train's binary_logloss: 0.681599 valid's binary_logloss: 0.687655 regularization_factors, val_score: 0.687553: 70%|####### | 14/20 [00:18<00:08, 1.38s/it][I 2020-09-27 05:01:39,213] Trial 56 finished with value: 0.6876553612371981 and parameters: {'lambda_l1': 0.08111866585830418, 'lambda_l2': 7.066527702724071e-06}. Best is trial 52 with value: 0.687553223456156. regularization_factors, val_score: 0.687553: 70%|####### | 14/20 [00:18<00:08, 1.38s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010624 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737 [200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967 [300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694 Early stopping, best iteration is: [231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553 regularization_factors, val_score: 0.687553: 75%|#######5 | 15/20 [00:19<00:06, 1.35s/it][I 2020-09-27 05:01:40,473] Trial 57 finished with value: 0.6875532234561976 and parameters: {'lambda_l1': 1.0664783935126406e-08, 'lambda_l2': 1.0078447980735048e-08}. Best is trial 52 with value: 0.687553223456156. regularization_factors, val_score: 0.687553: 75%|#######5 | 15/20 [00:19<00:06, 1.35s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007967 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737 [200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967 [300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694 Early stopping, best iteration is: [231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553 regularization_factors, val_score: 0.687553: 80%|######## | 16/20 [00:20<00:05, 1.32s/it][I 2020-09-27 05:01:41,748] Trial 58 finished with value: 0.6875532234661581 and parameters: {'lambda_l1': 1.2662033446419092e-08, 'lambda_l2': 1.5025969082548396e-06}. Best is trial 52 with value: 0.687553223456156. regularization_factors, val_score: 0.687553: 80%|######## | 16/20 [00:20<00:05, 1.32s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010265 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737 [200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967 [300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694 Early stopping, best iteration is: [231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553 regularization_factors, val_score: 0.687553: 85%|########5 | 17/20 [00:22<00:03, 1.30s/it][I 2020-09-27 05:01:43,006] Trial 59 finished with value: 0.6875532234563088 and parameters: {'lambda_l1': 1.9318545060845645e-08, 'lambda_l2': 1.3003225260249327e-08}. Best is trial 52 with value: 0.687553223456156. regularization_factors, val_score: 0.687553: 85%|########5 | 17/20 [00:22<00:03, 1.30s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008421 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737 [200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967 [300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694 Early stopping, best iteration is: [231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553 regularization_factors, val_score: 0.687553: 90%|######### | 18/20 [00:23<00:02, 1.28s/it][I 2020-09-27 05:01:44,230] Trial 60 finished with value: 0.6875532237690561 and parameters: {'lambda_l1': 1.4255011647002522e-05, 'lambda_l2': 1.2469961061897813e-06}. Best is trial 52 with value: 0.687553223456156. regularization_factors, val_score: 0.687553: 90%|######### | 18/20 [00:23<00:02, 1.28s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008041 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737 [200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967 [300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694 Early stopping, best iteration is: [231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553 regularization_factors, val_score: 0.687553: 95%|#########5| 19/20 [00:24<00:01, 1.27s/it][I 2020-09-27 05:01:45,477] Trial 61 finished with value: 0.6875532234563587 and parameters: {'lambda_l1': 1.5764030850824877e-08, 'lambda_l2': 1.4202912414449929e-08}. Best is trial 52 with value: 0.687553223456156. regularization_factors, val_score: 0.687553: 95%|#########5| 19/20 [00:24<00:01, 1.27s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007967 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737 [200] train's binary_logloss: 0.684545 valid's binary_logloss: 0.687967 [300] train's binary_logloss: 0.682519 valid's binary_logloss: 0.687694 Early stopping, best iteration is: [231] train's binary_logloss: 0.683858 valid's binary_logloss: 0.687553 regularization_factors, val_score: 0.687553: 100%|##########| 20/20 [00:25<00:00, 1.29s/it][I 2020-09-27 05:01:46,827] Trial 62 finished with value: 0.6875532234562169 and parameters: {'lambda_l1': 1.4154637690622022e-08, 'lambda_l2': 1.0757491754705414e-08}. Best is trial 52 with value: 0.687553223456156. regularization_factors, val_score: 0.687553: 100%|##########| 20/20 [00:25<00:00, 1.30s/it] min_data_in_leaf, val_score: 0.687553: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008157 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687147 valid's binary_logloss: 0.688925 [200] train's binary_logloss: 0.6848 valid's binary_logloss: 0.688284 [300] train's binary_logloss: 0.682838 valid's binary_logloss: 0.688287 Early stopping, best iteration is: [229] train's binary_logloss: 0.684208 valid's binary_logloss: 0.688185 min_data_in_leaf, val_score: 0.687553: 20%|## | 1/5 [00:01<00:05, 1.47s/it][I 2020-09-27 05:01:48,314] Trial 63 finished with value: 0.688184603164505 and parameters: {'min_child_samples': 100}. Best is trial 63 with value: 0.688184603164505. min_data_in_leaf, val_score: 0.687553: 20%|## | 1/5 [00:01<00:05, 1.47s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013040 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687067 valid's binary_logloss: 0.688767 [200] train's binary_logloss: 0.684549 valid's binary_logloss: 0.68803 [300] train's binary_logloss: 0.682536 valid's binary_logloss: 0.687853 Early stopping, best iteration is: [228] train's binary_logloss: 0.683958 valid's binary_logloss: 0.687779 min_data_in_leaf, val_score: 0.687553: 40%|#### | 2/5 [00:02<00:04, 1.43s/it][I 2020-09-27 05:01:49,650] Trial 64 finished with value: 0.6877794897053392 and parameters: {'min_child_samples': 10}. Best is trial 64 with value: 0.6877794897053392. min_data_in_leaf, val_score: 0.687553: 40%|#### | 2/5 [00:02<00:04, 1.43s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007873 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687067 valid's binary_logloss: 0.688767 [200] train's binary_logloss: 0.684546 valid's binary_logloss: 0.688205 [300] train's binary_logloss: 0.68251 valid's binary_logloss: 0.687847 [400] train's binary_logloss: 0.68055 valid's binary_logloss: 0.688135 Early stopping, best iteration is: [347] train's binary_logloss: 0.681552 valid's binary_logloss: 0.687815 min_data_in_leaf, val_score: 0.687553: 60%|###### | 3/5 [00:04<00:02, 1.50s/it][I 2020-09-27 05:01:51,310] Trial 65 finished with value: 0.6878151679233732 and parameters: {'min_child_samples': 5}. Best is trial 64 with value: 0.6877794897053392. min_data_in_leaf, val_score: 0.687553: 60%|###### | 3/5 [00:04<00:02, 1.50s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008336 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687068 valid's binary_logloss: 0.688737 [200] train's binary_logloss: 0.684557 valid's binary_logloss: 0.68783 [300] train's binary_logloss: 0.682493 valid's binary_logloss: 0.687685 Early stopping, best iteration is: [233] train's binary_logloss: 0.683814 valid's binary_logloss: 0.687524 min_data_in_leaf, val_score: 0.687524: 80%|######## | 4/5 [00:05<00:01, 1.42s/it][I 2020-09-27 05:01:52,557] Trial 66 finished with value: 0.6875238059643457 and parameters: {'min_child_samples': 25}. Best is trial 66 with value: 0.6875238059643457. min_data_in_leaf, val_score: 0.687524: 80%|######## | 4/5 [00:05<00:01, 1.42s/it][LightGBM] [Info] Number of positive: 46298, number of negative: 46728 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008121 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.497689 -> initscore=-0.009245 [LightGBM] [Info] Start training from score -0.009245 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687098 valid's binary_logloss: 0.688869 [200] train's binary_logloss: 0.684669 valid's binary_logloss: 0.688242 [300] train's binary_logloss: 0.68261 valid's binary_logloss: 0.688041 [400] train's binary_logloss: 0.680742 valid's binary_logloss: 0.688478 Early stopping, best iteration is: [302] train's binary_logloss: 0.682568 valid's binary_logloss: 0.688016 min_data_in_leaf, val_score: 0.687524: 100%|##########| 5/5 [00:07<00:00, 1.45s/it][I 2020-09-27 05:01:54,060] Trial 67 finished with value: 0.6880163054826545 and parameters: {'min_child_samples': 50}. Best is trial 66 with value: 0.6875238059643457. min_data_in_leaf, val_score: 0.687524: 100%|##########| 5/5 [00:07<00:00, 1.44s/it] Fold : 8 [I 2020-09-27 05:01:54,160] A new study created in memory with name: no-name-f83c045e-ae9d-45af-86f2-b4f88b2fc4c1 feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001841 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.6627 valid's binary_logloss: 0.6905 Early stopping, best iteration is: [38] train's binary_logloss: 0.677925 valid's binary_logloss: 0.689727 feature_fraction, val_score: 0.689727: 14%|#4 | 1/7 [00:01<00:06, 1.03s/it][I 2020-09-27 05:01:55,205] Trial 0 finished with value: 0.6897267085894525 and parameters: {'feature_fraction': 1.0}. Best is trial 0 with value: 0.6897267085894525. feature_fraction, val_score: 0.689727: 14%|#4 | 1/7 [00:01<00:06, 1.03s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000753 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.666345 valid's binary_logloss: 0.689793 Early stopping, best iteration is: [40] train's binary_logloss: 0.679261 valid's binary_logloss: 0.689685 feature_fraction, val_score: 0.689685: 29%|##8 | 2/7 [00:01<00:04, 1.07it/s][I 2020-09-27 05:01:55,922] Trial 1 finished with value: 0.689684519676036 and parameters: {'feature_fraction': 0.4}. Best is trial 1 with value: 0.689684519676036. feature_fraction, val_score: 0.689685: 29%|##8 | 2/7 [00:01<00:04, 1.07it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.033713 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664351 valid's binary_logloss: 0.689953 Early stopping, best iteration is: [82] train's binary_logloss: 0.668343 valid's binary_logloss: 0.689942 feature_fraction, val_score: 0.689685: 43%|####2 | 3/7 [00:02<00:03, 1.02it/s][I 2020-09-27 05:01:56,995] Trial 2 finished with value: 0.6899415616595388 and parameters: {'feature_fraction': 0.6}. Best is trial 1 with value: 0.689684519676036. feature_fraction, val_score: 0.689685: 43%|####2 | 3/7 [00:02<00:03, 1.02it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002095 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663723 valid's binary_logloss: 0.690154 Early stopping, best iteration is: [47] train's binary_logloss: 0.676135 valid's binary_logloss: 0.689446 feature_fraction, val_score: 0.689446: 57%|#####7 | 4/7 [00:03<00:02, 1.08it/s][I 2020-09-27 05:01:57,810] Trial 3 finished with value: 0.6894460749095064 and parameters: {'feature_fraction': 0.7}. Best is trial 3 with value: 0.6894460749095064. feature_fraction, val_score: 0.689446: 57%|#####7 | 4/7 [00:03<00:02, 1.08it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001658 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.662983 valid's binary_logloss: 0.690679 Early stopping, best iteration is: [53] train's binary_logloss: 0.674243 valid's binary_logloss: 0.689539 feature_fraction, val_score: 0.689446: 71%|#######1 | 5/7 [00:04<00:01, 1.10it/s][I 2020-09-27 05:01:58,677] Trial 4 finished with value: 0.6895387621340154 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 3 with value: 0.6894460749095064. feature_fraction, val_score: 0.689446: 71%|#######1 | 5/7 [00:04<00:01, 1.10it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008537 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663117 valid's binary_logloss: 0.689832 Early stopping, best iteration is: [47] train's binary_logloss: 0.675766 valid's binary_logloss: 0.689575 feature_fraction, val_score: 0.689446: 86%|########5 | 6/7 [00:05<00:00, 1.10it/s][I 2020-09-27 05:01:59,592] Trial 5 finished with value: 0.6895746649335026 and parameters: {'feature_fraction': 0.8}. Best is trial 3 with value: 0.6894460749095064. feature_fraction, val_score: 0.689446: 86%|########5 | 6/7 [00:05<00:00, 1.10it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000930 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665484 valid's binary_logloss: 0.690502 Early stopping, best iteration is: [82] train's binary_logloss: 0.669205 valid's binary_logloss: 0.69007 feature_fraction, val_score: 0.689446: 100%|##########| 7/7 [00:06<00:00, 1.09it/s][I 2020-09-27 05:02:00,511] Trial 6 finished with value: 0.6900695949496937 and parameters: {'feature_fraction': 0.5}. Best is trial 3 with value: 0.6894460749095064. feature_fraction, val_score: 0.689446: 100%|##########| 7/7 [00:06<00:00, 1.10it/s] num_leaves, val_score: 0.689446: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008662 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.686038 valid's binary_logloss: 0.689757 [200] train's binary_logloss: 0.68281 valid's binary_logloss: 0.688899 [300] train's binary_logloss: 0.68005 valid's binary_logloss: 0.688585 [400] train's binary_logloss: 0.677584 valid's binary_logloss: 0.688509 Early stopping, best iteration is: [336] train's binary_logloss: 0.679139 valid's binary_logloss: 0.688453 num_leaves, val_score: 0.688453: 5%|5 | 1/20 [00:01<00:30, 1.59s/it][I 2020-09-27 05:02:02,118] Trial 7 finished with value: 0.6884528114970678 and parameters: {'num_leaves': 5}. Best is trial 7 with value: 0.6884528114970678. num_leaves, val_score: 0.688453: 5%|5 | 1/20 [00:01<00:30, 1.59s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012715 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.593349 valid's binary_logloss: 0.694762 Early stopping, best iteration is: [28] train's binary_logloss: 0.655278 valid's binary_logloss: 0.690329 num_leaves, val_score: 0.688453: 10%|# | 2/20 [00:02<00:25, 1.43s/it][I 2020-09-27 05:02:03,178] Trial 8 finished with value: 0.690328689997322 and parameters: {'num_leaves': 139}. Best is trial 7 with value: 0.6884528114970678. num_leaves, val_score: 0.688453: 10%|# | 2/20 [00:02<00:25, 1.43s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012613 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.545663 valid's binary_logloss: 0.696241 Early stopping, best iteration is: [18] train's binary_logloss: 0.654494 valid's binary_logloss: 0.690387 num_leaves, val_score: 0.688453: 15%|#5 | 3/20 [00:03<00:23, 1.37s/it][I 2020-09-27 05:02:04,416] Trial 9 finished with value: 0.6903874648728597 and parameters: {'num_leaves': 227}. Best is trial 7 with value: 0.6884528114970678. num_leaves, val_score: 0.688453: 15%|#5 | 3/20 [00:03<00:23, 1.37s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012246 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689978 valid's binary_logloss: 0.691198 [200] train's binary_logloss: 0.688797 valid's binary_logloss: 0.69055 [300] train's binary_logloss: 0.688072 valid's binary_logloss: 0.690072 [400] train's binary_logloss: 0.687556 valid's binary_logloss: 0.689833 [500] train's binary_logloss: 0.687169 valid's binary_logloss: 0.689688 [600] train's binary_logloss: 0.686868 valid's binary_logloss: 0.689597 [700] train's binary_logloss: 0.686624 valid's binary_logloss: 0.689542 [800] train's binary_logloss: 0.68642 valid's binary_logloss: 0.689474 [900] train's binary_logloss: 0.686246 valid's binary_logloss: 0.689462 [1000] train's binary_logloss: 0.686095 valid's binary_logloss: 0.689411 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.686095 valid's binary_logloss: 0.689411 num_leaves, val_score: 0.688453: 20%|## | 4/20 [00:06<00:30, 1.89s/it][I 2020-09-27 05:02:07,521] Trial 10 finished with value: 0.6894108839720557 and parameters: {'num_leaves': 2}. Best is trial 7 with value: 0.6884528114970678. num_leaves, val_score: 0.688453: 20%|## | 4/20 [00:06<00:30, 1.89s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008181 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.684097 valid's binary_logloss: 0.689543 [200] train's binary_logloss: 0.679346 valid's binary_logloss: 0.689259 [300] train's binary_logloss: 0.675305 valid's binary_logloss: 0.68922 [400] train's binary_logloss: 0.671448 valid's binary_logloss: 0.689174 Early stopping, best iteration is: [385] train's binary_logloss: 0.672024 valid's binary_logloss: 0.689088 num_leaves, val_score: 0.688453: 25%|##5 | 5/20 [00:08<00:28, 1.88s/it][I 2020-09-27 05:02:09,364] Trial 11 finished with value: 0.6890882300993826 and parameters: {'num_leaves': 7}. Best is trial 7 with value: 0.6884528114970678. num_leaves, val_score: 0.688453: 25%|##5 | 5/20 [00:08<00:28, 1.88s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009553 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.684097 valid's binary_logloss: 0.689543 [200] train's binary_logloss: 0.679346 valid's binary_logloss: 0.689259 [300] train's binary_logloss: 0.675305 valid's binary_logloss: 0.68922 [400] train's binary_logloss: 0.671448 valid's binary_logloss: 0.689174 Early stopping, best iteration is: [385] train's binary_logloss: 0.672024 valid's binary_logloss: 0.689088 num_leaves, val_score: 0.688453: 30%|### | 6/20 [00:10<00:25, 1.85s/it][I 2020-09-27 05:02:11,153] Trial 12 finished with value: 0.6890882300993826 and parameters: {'num_leaves': 7}. Best is trial 7 with value: 0.6884528114970678. num_leaves, val_score: 0.688453: 30%|### | 6/20 [00:10<00:25, 1.85s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008338 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.652737 valid's binary_logloss: 0.690159 Early stopping, best iteration is: [30] train's binary_logloss: 0.676594 valid's binary_logloss: 0.689592 num_leaves, val_score: 0.688453: 35%|###5 | 7/20 [00:11<00:19, 1.52s/it][I 2020-09-27 05:02:11,891] Trial 13 finished with value: 0.6895920137463518 and parameters: {'num_leaves': 46}. Best is trial 7 with value: 0.6884528114970678. num_leaves, val_score: 0.688453: 35%|###5 | 7/20 [00:11<00:19, 1.52s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008791 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.630115 valid's binary_logloss: 0.692022 Early stopping, best iteration is: [25] train's binary_logloss: 0.671325 valid's binary_logloss: 0.690619 num_leaves, val_score: 0.688453: 40%|#### | 8/20 [00:12<00:15, 1.33s/it][I 2020-09-27 05:02:12,774] Trial 14 finished with value: 0.690619286117565 and parameters: {'num_leaves': 79}. Best is trial 7 with value: 0.6884528114970678. num_leaves, val_score: 0.688453: 40%|#### | 8/20 [00:12<00:15, 1.33s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008004 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.614161 valid's binary_logloss: 0.692813 Early stopping, best iteration is: [31] train's binary_logloss: 0.661053 valid's binary_logloss: 0.690497 num_leaves, val_score: 0.688453: 45%|####5 | 9/20 [00:13<00:13, 1.21s/it][I 2020-09-27 05:02:13,713] Trial 15 finished with value: 0.6904972906779976 and parameters: {'num_leaves': 104}. Best is trial 7 with value: 0.6884528114970678. num_leaves, val_score: 0.688453: 45%|####5 | 9/20 [00:13<00:13, 1.21s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008054 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685119 valid's binary_logloss: 0.68988 [200] train's binary_logloss: 0.681155 valid's binary_logloss: 0.689507 Early stopping, best iteration is: [189] train's binary_logloss: 0.681535 valid's binary_logloss: 0.689459 num_leaves, val_score: 0.688453: 50%|##### | 10/20 [00:14<00:11, 1.17s/it][I 2020-09-27 05:02:14,772] Trial 16 finished with value: 0.6894592704386628 and parameters: {'num_leaves': 6}. Best is trial 7 with value: 0.6884528114970678. num_leaves, val_score: 0.688453: 50%|##### | 10/20 [00:14<00:11, 1.17s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009175 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.568944 valid's binary_logloss: 0.695405 Early stopping, best iteration is: [15] train's binary_logloss: 0.665473 valid's binary_logloss: 0.690687 num_leaves, val_score: 0.688453: 55%|#####5 | 11/20 [00:15<00:10, 1.13s/it][I 2020-09-27 05:02:15,813] Trial 17 finished with value: 0.6906872291686416 and parameters: {'num_leaves': 182}. Best is trial 7 with value: 0.6884528114970678. num_leaves, val_score: 0.688453: 55%|#####5 | 11/20 [00:15<00:10, 1.13s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001779 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.649856 valid's binary_logloss: 0.690204 Early stopping, best iteration is: [48] train's binary_logloss: 0.667974 valid's binary_logloss: 0.689828 num_leaves, val_score: 0.688453: 60%|###### | 12/20 [00:16<00:08, 1.06s/it][I 2020-09-27 05:02:16,731] Trial 18 finished with value: 0.6898284619368371 and parameters: {'num_leaves': 50}. Best is trial 7 with value: 0.6884528114970678. num_leaves, val_score: 0.688453: 60%|###### | 12/20 [00:16<00:08, 1.06s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009546 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.652737 valid's binary_logloss: 0.690159 Early stopping, best iteration is: [30] train's binary_logloss: 0.676594 valid's binary_logloss: 0.689592 num_leaves, val_score: 0.688453: 65%|######5 | 13/20 [00:16<00:06, 1.03it/s][I 2020-09-27 05:02:17,476] Trial 19 finished with value: 0.6895920137463518 and parameters: {'num_leaves': 46}. Best is trial 7 with value: 0.6884528114970678. num_leaves, val_score: 0.688453: 65%|######5 | 13/20 [00:16<00:06, 1.03it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011684 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.688384 valid's binary_logloss: 0.6906 [200] train's binary_logloss: 0.686574 valid's binary_logloss: 0.689853 [300] train's binary_logloss: 0.68527 valid's binary_logloss: 0.689427 [400] train's binary_logloss: 0.684182 valid's binary_logloss: 0.689235 [500] train's binary_logloss: 0.683198 valid's binary_logloss: 0.689083 [600] train's binary_logloss: 0.6823 valid's binary_logloss: 0.688973 [700] train's binary_logloss: 0.681468 valid's binary_logloss: 0.688875 [800] train's binary_logloss: 0.680639 valid's binary_logloss: 0.688913 Early stopping, best iteration is: [703] train's binary_logloss: 0.681443 valid's binary_logloss: 0.68886 num_leaves, val_score: 0.688453: 70%|####### | 14/20 [00:19<00:08, 1.48s/it][I 2020-09-27 05:02:20,147] Trial 20 finished with value: 0.6888601909760304 and parameters: {'num_leaves': 3}. Best is trial 7 with value: 0.6884528114970678. num_leaves, val_score: 0.688453: 70%|####### | 14/20 [00:19<00:08, 1.48s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008648 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.676993 valid's binary_logloss: 0.689308 Early stopping, best iteration is: [90] train's binary_logloss: 0.678106 valid's binary_logloss: 0.689282 num_leaves, val_score: 0.688453: 75%|#######5 | 15/20 [00:20<00:06, 1.28s/it][I 2020-09-27 05:02:20,976] Trial 21 finished with value: 0.6892821753767159 and parameters: {'num_leaves': 15}. Best is trial 7 with value: 0.6884528114970678. num_leaves, val_score: 0.688453: 75%|#######5 | 15/20 [00:20<00:06, 1.28s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008704 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.689978 valid's binary_logloss: 0.691198 [200] train's binary_logloss: 0.688797 valid's binary_logloss: 0.69055 [300] train's binary_logloss: 0.688072 valid's binary_logloss: 0.690072 [400] train's binary_logloss: 0.687556 valid's binary_logloss: 0.689833 [500] train's binary_logloss: 0.687169 valid's binary_logloss: 0.689688 [600] train's binary_logloss: 0.686868 valid's binary_logloss: 0.689597 [700] train's binary_logloss: 0.686624 valid's binary_logloss: 0.689542 [800] train's binary_logloss: 0.68642 valid's binary_logloss: 0.689474 [900] train's binary_logloss: 0.686246 valid's binary_logloss: 0.689462 [1000] train's binary_logloss: 0.686095 valid's binary_logloss: 0.689411 Did not meet early stopping. Best iteration is: [1000] train's binary_logloss: 0.686095 valid's binary_logloss: 0.689411 num_leaves, val_score: 0.688453: 80%|######## | 16/20 [00:25<00:09, 2.43s/it][I 2020-09-27 05:02:26,077] Trial 22 finished with value: 0.6894108839720557 and parameters: {'num_leaves': 2}. Best is trial 7 with value: 0.6884528114970678. num_leaves, val_score: 0.688453: 80%|######## | 16/20 [00:25<00:09, 2.43s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.016684 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.65797 valid's binary_logloss: 0.690213 Early stopping, best iteration is: [38] train's binary_logloss: 0.675789 valid's binary_logloss: 0.689489 num_leaves, val_score: 0.688453: 85%|########5 | 17/20 [00:26<00:05, 1.96s/it][I 2020-09-27 05:02:26,941] Trial 23 finished with value: 0.6894891049398579 and parameters: {'num_leaves': 39}. Best is trial 7 with value: 0.6884528114970678. num_leaves, val_score: 0.688453: 85%|########5 | 17/20 [00:26<00:05, 1.96s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011681 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.632096 valid's binary_logloss: 0.691354 Early stopping, best iteration is: [28] train's binary_logloss: 0.670154 valid's binary_logloss: 0.690155 num_leaves, val_score: 0.688453: 90%|######### | 18/20 [00:27<00:03, 1.64s/it][I 2020-09-27 05:02:27,818] Trial 24 finished with value: 0.6901553410011205 and parameters: {'num_leaves': 75}. Best is trial 7 with value: 0.6884528114970678. num_leaves, val_score: 0.688453: 90%|######### | 18/20 [00:27<00:03, 1.64s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012016 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.669171 valid's binary_logloss: 0.689591 Early stopping, best iteration is: [78] train's binary_logloss: 0.67319 valid's binary_logloss: 0.68921 num_leaves, val_score: 0.688453: 95%|#########5| 19/20 [00:28<00:01, 1.42s/it][I 2020-09-27 05:02:28,738] Trial 25 finished with value: 0.6892103907051438 and parameters: {'num_leaves': 24}. Best is trial 7 with value: 0.6884528114970678. num_leaves, val_score: 0.688453: 95%|#########5| 19/20 [00:28<00:01, 1.42s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.018748 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.631577 valid's binary_logloss: 0.692299 Early stopping, best iteration is: [26] train's binary_logloss: 0.671068 valid's binary_logloss: 0.690199 num_leaves, val_score: 0.688453: 100%|##########| 20/20 [00:29<00:00, 1.26s/it][I 2020-09-27 05:02:29,625] Trial 26 finished with value: 0.6901985363739374 and parameters: {'num_leaves': 77}. Best is trial 7 with value: 0.6884528114970678. num_leaves, val_score: 0.688453: 100%|##########| 20/20 [00:29<00:00, 1.46s/it] bagging, val_score: 0.688453: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001613 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.686006 valid's binary_logloss: 0.689628 [200] train's binary_logloss: 0.682779 valid's binary_logloss: 0.688355 [300] train's binary_logloss: 0.68005 valid's binary_logloss: 0.688419 Early stopping, best iteration is: [219] train's binary_logloss: 0.682239 valid's binary_logloss: 0.688258 bagging, val_score: 0.688258: 10%|# | 1/10 [00:01<00:11, 1.23s/it][I 2020-09-27 05:02:30,868] Trial 27 finished with value: 0.6882582742878182 and parameters: {'bagging_fraction': 0.615338737094182, 'bagging_freq': 7}. Best is trial 27 with value: 0.6882582742878182. bagging, val_score: 0.688258: 10%|# | 1/10 [00:01<00:11, 1.23s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008365 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68604 valid's binary_logloss: 0.689719 [200] train's binary_logloss: 0.682712 valid's binary_logloss: 0.68842 Early stopping, best iteration is: [194] train's binary_logloss: 0.682897 valid's binary_logloss: 0.688374 bagging, val_score: 0.688258: 20%|## | 2/10 [00:02<00:09, 1.19s/it][I 2020-09-27 05:02:31,952] Trial 28 finished with value: 0.6883743964950245 and parameters: {'bagging_fraction': 0.6026436236346611, 'bagging_freq': 7}. Best is trial 27 with value: 0.6882582742878182. bagging, val_score: 0.688258: 20%|## | 2/10 [00:02<00:09, 1.19s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008715 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.686015 valid's binary_logloss: 0.689456 [200] train's binary_logloss: 0.682706 valid's binary_logloss: 0.688437 [300] train's binary_logloss: 0.67996 valid's binary_logloss: 0.688695 Early stopping, best iteration is: [223] train's binary_logloss: 0.682071 valid's binary_logloss: 0.68839 bagging, val_score: 0.688258: 30%|### | 3/10 [00:03<00:08, 1.19s/it][I 2020-09-27 05:02:33,149] Trial 29 finished with value: 0.6883895614361045 and parameters: {'bagging_fraction': 0.6029377567256127, 'bagging_freq': 7}. Best is trial 27 with value: 0.6882582742878182. bagging, val_score: 0.688258: 30%|### | 3/10 [00:03<00:08, 1.19s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008131 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685948 valid's binary_logloss: 0.689692 [200] train's binary_logloss: 0.68277 valid's binary_logloss: 0.688533 [300] train's binary_logloss: 0.679958 valid's binary_logloss: 0.688725 Early stopping, best iteration is: [274] train's binary_logloss: 0.680722 valid's binary_logloss: 0.688429 bagging, val_score: 0.688258: 40%|#### | 4/10 [00:04<00:07, 1.22s/it][I 2020-09-27 05:02:34,457] Trial 30 finished with value: 0.6884290873607732 and parameters: {'bagging_fraction': 0.5938506890500748, 'bagging_freq': 7}. Best is trial 27 with value: 0.6882582742878182. bagging, val_score: 0.688258: 40%|#### | 4/10 [00:04<00:07, 1.22s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013062 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685972 valid's binary_logloss: 0.689581 [200] train's binary_logloss: 0.682787 valid's binary_logloss: 0.688637 Early stopping, best iteration is: [196] train's binary_logloss: 0.682904 valid's binary_logloss: 0.688622 bagging, val_score: 0.688258: 50%|##### | 5/10 [00:05<00:05, 1.18s/it][I 2020-09-27 05:02:35,525] Trial 31 finished with value: 0.6886216049092391 and parameters: {'bagging_fraction': 0.5957530118583011, 'bagging_freq': 7}. Best is trial 27 with value: 0.6882582742878182. bagging, val_score: 0.688258: 50%|##### | 5/10 [00:05<00:05, 1.18s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008942 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.686042 valid's binary_logloss: 0.68983 [200] train's binary_logloss: 0.682839 valid's binary_logloss: 0.688484 Early stopping, best iteration is: [196] train's binary_logloss: 0.682969 valid's binary_logloss: 0.688429 bagging, val_score: 0.688258: 60%|###### | 6/10 [00:06<00:04, 1.15s/it][I 2020-09-27 05:02:36,622] Trial 32 finished with value: 0.6884291730078321 and parameters: {'bagging_fraction': 0.601131517375276, 'bagging_freq': 7}. Best is trial 27 with value: 0.6882582742878182. bagging, val_score: 0.688258: 60%|###### | 6/10 [00:06<00:04, 1.15s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008379 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685889 valid's binary_logloss: 0.68974 [200] train's binary_logloss: 0.682644 valid's binary_logloss: 0.68847 [300] train's binary_logloss: 0.679926 valid's binary_logloss: 0.688667 Early stopping, best iteration is: [200] train's binary_logloss: 0.682644 valid's binary_logloss: 0.68847 bagging, val_score: 0.688258: 70%|####### | 7/10 [00:08<00:03, 1.14s/it][I 2020-09-27 05:02:37,721] Trial 33 finished with value: 0.6884699706432157 and parameters: {'bagging_fraction': 0.6161422919690354, 'bagging_freq': 7}. Best is trial 27 with value: 0.6882582742878182. bagging, val_score: 0.688258: 70%|####### | 7/10 [00:08<00:03, 1.14s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013266 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.686069 valid's binary_logloss: 0.689747 [200] train's binary_logloss: 0.682773 valid's binary_logloss: 0.688633 [300] train's binary_logloss: 0.679937 valid's binary_logloss: 0.68895 Early stopping, best iteration is: [205] train's binary_logloss: 0.682645 valid's binary_logloss: 0.688588 bagging, val_score: 0.688258: 80%|######## | 8/10 [00:09<00:02, 1.14s/it][I 2020-09-27 05:02:38,853] Trial 34 finished with value: 0.6885878332172652 and parameters: {'bagging_fraction': 0.6004490251930547, 'bagging_freq': 7}. Best is trial 27 with value: 0.6882582742878182. bagging, val_score: 0.688258: 80%|######## | 8/10 [00:09<00:02, 1.14s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008258 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.686065 valid's binary_logloss: 0.689456 [200] train's binary_logloss: 0.682802 valid's binary_logloss: 0.688959 [300] train's binary_logloss: 0.680026 valid's binary_logloss: 0.688847 Early stopping, best iteration is: [262] train's binary_logloss: 0.681101 valid's binary_logloss: 0.688619 bagging, val_score: 0.688258: 90%|######### | 9/10 [00:10<00:01, 1.21s/it][I 2020-09-27 05:02:40,228] Trial 35 finished with value: 0.6886193436195897 and parameters: {'bagging_fraction': 0.7777960275374113, 'bagging_freq': 7}. Best is trial 27 with value: 0.6882582742878182. bagging, val_score: 0.688258: 90%|######### | 9/10 [00:10<00:01, 1.21s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008496 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685807 valid's binary_logloss: 0.689065 [200] train's binary_logloss: 0.68264 valid's binary_logloss: 0.6887 [300] train's binary_logloss: 0.680006 valid's binary_logloss: 0.688483 Early stopping, best iteration is: [291] train's binary_logloss: 0.680267 valid's binary_logloss: 0.688364 bagging, val_score: 0.688258: 100%|##########| 10/10 [00:11<00:00, 1.25s/it][I 2020-09-27 05:02:41,562] Trial 36 finished with value: 0.6883635702761751 and parameters: {'bagging_fraction': 0.4580982361208543, 'bagging_freq': 4}. Best is trial 27 with value: 0.6882582742878182. bagging, val_score: 0.688258: 100%|##########| 10/10 [00:11<00:00, 1.19s/it] feature_fraction_stage2, val_score: 0.688258: 0%| | 0/6 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008209 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486 [200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315 [300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232 [400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467 Early stopping, best iteration is: [334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072 feature_fraction_stage2, val_score: 0.688072: 17%|#6 | 1/6 [00:01<00:07, 1.55s/it][I 2020-09-27 05:02:43,127] Trial 37 finished with value: 0.6880722259887746 and parameters: {'feature_fraction': 0.748}. Best is trial 37 with value: 0.6880722259887746. feature_fraction_stage2, val_score: 0.688072: 17%|#6 | 1/6 [00:01<00:07, 1.55s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.003567 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486 [200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315 [300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232 [400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467 Early stopping, best iteration is: [334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072 feature_fraction_stage2, val_score: 0.688072: 33%|###3 | 2/6 [00:03<00:06, 1.58s/it][I 2020-09-27 05:02:44,774] Trial 38 finished with value: 0.6880722259887746 and parameters: {'feature_fraction': 0.716}. Best is trial 37 with value: 0.6880722259887746. feature_fraction_stage2, val_score: 0.688072: 33%|###3 | 2/6 [00:03<00:06, 1.58s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008780 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.686006 valid's binary_logloss: 0.689628 [200] train's binary_logloss: 0.682779 valid's binary_logloss: 0.688355 [300] train's binary_logloss: 0.68005 valid's binary_logloss: 0.688419 Early stopping, best iteration is: [219] train's binary_logloss: 0.682239 valid's binary_logloss: 0.688258 feature_fraction_stage2, val_score: 0.688072: 50%|##### | 3/6 [00:04<00:04, 1.45s/it][I 2020-09-27 05:02:45,923] Trial 39 finished with value: 0.6882582742878182 and parameters: {'feature_fraction': 0.6839999999999999}. Best is trial 37 with value: 0.6880722259887746. feature_fraction_stage2, val_score: 0.688072: 50%|##### | 3/6 [00:04<00:04, 1.45s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009262 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685951 valid's binary_logloss: 0.689912 [200] train's binary_logloss: 0.682613 valid's binary_logloss: 0.688642 [300] train's binary_logloss: 0.679824 valid's binary_logloss: 0.688952 Early stopping, best iteration is: [201] train's binary_logloss: 0.682582 valid's binary_logloss: 0.688618 feature_fraction_stage2, val_score: 0.688072: 67%|######6 | 4/6 [00:05<00:02, 1.36s/it][I 2020-09-27 05:02:47,065] Trial 40 finished with value: 0.6886180547556369 and parameters: {'feature_fraction': 0.7799999999999999}. Best is trial 37 with value: 0.6880722259887746. feature_fraction_stage2, val_score: 0.688072: 67%|######6 | 4/6 [00:05<00:02, 1.36s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009195 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685997 valid's binary_logloss: 0.689645 [200] train's binary_logloss: 0.682772 valid's binary_logloss: 0.688789 [300] train's binary_logloss: 0.680042 valid's binary_logloss: 0.689271 Early stopping, best iteration is: [200] train's binary_logloss: 0.682772 valid's binary_logloss: 0.688789 feature_fraction_stage2, val_score: 0.688072: 83%|########3 | 5/6 [00:06<00:01, 1.29s/it][I 2020-09-27 05:02:48,184] Trial 41 finished with value: 0.6887892049667754 and parameters: {'feature_fraction': 0.62}. Best is trial 37 with value: 0.6880722259887746. feature_fraction_stage2, val_score: 0.688072: 83%|########3 | 5/6 [00:06<00:01, 1.29s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013986 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.686063 valid's binary_logloss: 0.68975 [200] train's binary_logloss: 0.682823 valid's binary_logloss: 0.688614 [300] train's binary_logloss: 0.680059 valid's binary_logloss: 0.689073 Early stopping, best iteration is: [206] train's binary_logloss: 0.68266 valid's binary_logloss: 0.688579 feature_fraction_stage2, val_score: 0.688072: 100%|##########| 6/6 [00:07<00:00, 1.24s/it][I 2020-09-27 05:02:49,309] Trial 42 finished with value: 0.688578931631993 and parameters: {'feature_fraction': 0.652}. Best is trial 37 with value: 0.6880722259887746. feature_fraction_stage2, val_score: 0.688072: 100%|##########| 6/6 [00:07<00:00, 1.29s/it] regularization_factors, val_score: 0.688072: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001438 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.686053 valid's binary_logloss: 0.689655 [200] train's binary_logloss: 0.682817 valid's binary_logloss: 0.688251 [300] train's binary_logloss: 0.680044 valid's binary_logloss: 0.688312 Early stopping, best iteration is: [274] train's binary_logloss: 0.68083 valid's binary_logloss: 0.68815 regularization_factors, val_score: 0.688072: 5%|5 | 1/20 [00:01<00:26, 1.40s/it][I 2020-09-27 05:02:50,722] Trial 43 finished with value: 0.6881495198006506 and parameters: {'lambda_l1': 0.0485944386213443, 'lambda_l2': 0.24935601021074602}. Best is trial 43 with value: 0.6881495198006506. regularization_factors, val_score: 0.688072: 5%|5 | 1/20 [00:01<00:26, 1.40s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001480 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68603 valid's binary_logloss: 0.689695 [200] train's binary_logloss: 0.682787 valid's binary_logloss: 0.688424 [300] train's binary_logloss: 0.680205 valid's binary_logloss: 0.688656 Early stopping, best iteration is: [221] train's binary_logloss: 0.682216 valid's binary_logloss: 0.688392 regularization_factors, val_score: 0.688072: 10%|# | 2/20 [00:02<00:24, 1.35s/it][I 2020-09-27 05:02:51,960] Trial 44 finished with value: 0.6883919982325207 and parameters: {'lambda_l1': 0.10751881913962534, 'lambda_l2': 0.5824729388634683}. Best is trial 43 with value: 0.6881495198006506. regularization_factors, val_score: 0.688072: 10%|# | 2/20 [00:02<00:24, 1.35s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008830 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689487 [200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315 [300] train's binary_logloss: 0.679962 valid's binary_logloss: 0.688232 [400] train's binary_logloss: 0.677326 valid's binary_logloss: 0.688466 Early stopping, best iteration is: [334] train's binary_logloss: 0.679068 valid's binary_logloss: 0.688072 regularization_factors, val_score: 0.688072: 15%|#5 | 3/20 [00:04<00:24, 1.41s/it][I 2020-09-27 05:02:53,524] Trial 45 finished with value: 0.6880722890783393 and parameters: {'lambda_l1': 1.188253884167047e-06, 'lambda_l2': 0.005588276816712126}. Best is trial 45 with value: 0.6880722890783393. regularization_factors, val_score: 0.688072: 15%|#5 | 3/20 [00:04<00:24, 1.41s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002086 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689487 [200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315 [300] train's binary_logloss: 0.679962 valid's binary_logloss: 0.688232 [400] train's binary_logloss: 0.677326 valid's binary_logloss: 0.688466 Early stopping, best iteration is: [334] train's binary_logloss: 0.679068 valid's binary_logloss: 0.688072 regularization_factors, val_score: 0.688072: 20%|## | 4/20 [00:05<00:23, 1.48s/it][I 2020-09-27 05:02:55,155] Trial 46 finished with value: 0.6880722851297056 and parameters: {'lambda_l1': 5.105795902683296e-08, 'lambda_l2': 0.005237582202723369}. Best is trial 46 with value: 0.6880722851297056. regularization_factors, val_score: 0.688072: 20%|## | 4/20 [00:05<00:23, 1.48s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001606 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689487 [200] train's binary_logloss: 0.682685 valid's binary_logloss: 0.688315 [300] train's binary_logloss: 0.679962 valid's binary_logloss: 0.688232 [400] train's binary_logloss: 0.67738 valid's binary_logloss: 0.68864 Early stopping, best iteration is: [330] train's binary_logloss: 0.679165 valid's binary_logloss: 0.688085 regularization_factors, val_score: 0.688072: 25%|##5 | 5/20 [00:07<00:22, 1.52s/it][I 2020-09-27 05:02:56,785] Trial 47 finished with value: 0.6880848414842485 and parameters: {'lambda_l1': 1.720232956611629e-08, 'lambda_l2': 0.007516638600538305}. Best is trial 46 with value: 0.6880722851297056. regularization_factors, val_score: 0.688072: 25%|##5 | 5/20 [00:07<00:22, 1.52s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.014048 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486 [200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315 [300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232 [400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467 Early stopping, best iteration is: [334] train's binary_logloss: 0.679067 valid's binary_logloss: 0.688072 regularization_factors, val_score: 0.688072: 30%|### | 6/20 [00:09<00:21, 1.54s/it][I 2020-09-27 05:02:58,355] Trial 48 finished with value: 0.688072230112268 and parameters: {'lambda_l1': 1.027357618260445e-08, 'lambda_l2': 0.00036547314253742376}. Best is trial 48 with value: 0.688072230112268. regularization_factors, val_score: 0.688072: 30%|### | 6/20 [00:09<00:21, 1.54s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008658 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486 [200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315 [300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232 [400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467 Early stopping, best iteration is: [334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072 regularization_factors, val_score: 0.688072: 35%|###5 | 7/20 [00:10<00:20, 1.55s/it][I 2020-09-27 05:02:59,927] Trial 49 finished with value: 0.6880722260400305 and parameters: {'lambda_l1': 2.5990203700513686e-08, 'lambda_l2': 4.6609028631019925e-06}. Best is trial 49 with value: 0.6880722260400305. regularization_factors, val_score: 0.688072: 35%|###5 | 7/20 [00:10<00:20, 1.55s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001464 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486 [200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315 [300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232 [400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467 Early stopping, best iteration is: [334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072 regularization_factors, val_score: 0.688072: 40%|#### | 8/20 [00:12<00:18, 1.56s/it][I 2020-09-27 05:03:01,518] Trial 50 finished with value: 0.6880722259938851 and parameters: {'lambda_l1': 1.0501739122696127e-08, 'lambda_l2': 2.906436684517273e-07}. Best is trial 50 with value: 0.6880722259938851. regularization_factors, val_score: 0.688072: 40%|#### | 8/20 [00:12<00:18, 1.56s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008492 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486 [200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315 [300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232 [400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467 Early stopping, best iteration is: [334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072 regularization_factors, val_score: 0.688072: 45%|####5 | 9/20 [00:13<00:17, 1.56s/it][I 2020-09-27 05:03:03,084] Trial 51 finished with value: 0.6880722259917854 and parameters: {'lambda_l1': 1.0110083939507621e-08, 'lambda_l2': 1.8869898459706898e-07}. Best is trial 51 with value: 0.6880722259917854. regularization_factors, val_score: 0.688072: 45%|####5 | 9/20 [00:13<00:17, 1.56s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008606 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486 [200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315 [300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232 [400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467 Early stopping, best iteration is: [334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072 regularization_factors, val_score: 0.688072: 50%|##### | 10/20 [00:15<00:15, 1.56s/it][I 2020-09-27 05:03:04,652] Trial 52 finished with value: 0.6880722259908346 and parameters: {'lambda_l1': 1.5445331505275187e-08, 'lambda_l2': 1.2613210373316155e-07}. Best is trial 52 with value: 0.6880722259908346. regularization_factors, val_score: 0.688072: 50%|##### | 10/20 [00:15<00:15, 1.56s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008206 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486 [200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315 [300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232 [400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467 Early stopping, best iteration is: [334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072 regularization_factors, val_score: 0.688072: 55%|#####5 | 11/20 [00:16<00:14, 1.56s/it][I 2020-09-27 05:03:06,216] Trial 53 finished with value: 0.6880722259908322 and parameters: {'lambda_l1': 1.187725756223345e-08, 'lambda_l2': 8.260605690107965e-08}. Best is trial 53 with value: 0.6880722259908322. regularization_factors, val_score: 0.688072: 55%|#####5 | 11/20 [00:16<00:14, 1.56s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001731 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486 [200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315 [300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232 [400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467 Early stopping, best iteration is: [334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072 regularization_factors, val_score: 0.688072: 60%|###### | 12/20 [00:18<00:12, 1.59s/it][I 2020-09-27 05:03:07,867] Trial 54 finished with value: 0.6880722259624853 and parameters: {'lambda_l1': 2.1107731769847976e-06, 'lambda_l2': 1.024124993705147e-08}. Best is trial 54 with value: 0.6880722259624853. regularization_factors, val_score: 0.688072: 60%|###### | 12/20 [00:18<00:12, 1.59s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013514 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486 [200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315 [300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232 [400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467 Early stopping, best iteration is: [334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072 regularization_factors, val_score: 0.688072: 65%|######5 | 13/20 [00:20<00:11, 1.59s/it][I 2020-09-27 05:03:09,440] Trial 55 finished with value: 0.6880722259228028 and parameters: {'lambda_l1': 4.841910937380013e-06, 'lambda_l2': 1.6262995524915506e-08}. Best is trial 55 with value: 0.6880722259228028. regularization_factors, val_score: 0.688072: 65%|######5 | 13/20 [00:20<00:11, 1.59s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.002085 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486 [200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315 [300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232 [400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467 Early stopping, best iteration is: [334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072 regularization_factors, val_score: 0.688072: 70%|####### | 14/20 [00:21<00:09, 1.60s/it][I 2020-09-27 05:03:11,073] Trial 56 finished with value: 0.6880722257360149 and parameters: {'lambda_l1': 1.964871251332563e-05, 'lambda_l2': 1.3890624113900224e-08}. Best is trial 56 with value: 0.6880722257360149. regularization_factors, val_score: 0.688072: 70%|####### | 14/20 [00:21<00:09, 1.60s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013271 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486 [200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315 [300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232 [400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467 Early stopping, best iteration is: [334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072 regularization_factors, val_score: 0.688072: 75%|#######5 | 15/20 [00:23<00:07, 1.59s/it][I 2020-09-27 05:03:12,639] Trial 57 finished with value: 0.6880722255253509 and parameters: {'lambda_l1': 3.358919485655909e-05, 'lambda_l2': 1.6266181828610483e-08}. Best is trial 57 with value: 0.6880722255253509. regularization_factors, val_score: 0.688072: 75%|#######5 | 15/20 [00:23<00:07, 1.59s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009828 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486 [200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315 [300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232 [400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467 Early stopping, best iteration is: [334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072 regularization_factors, val_score: 0.688072: 80%|######## | 16/20 [00:24<00:06, 1.58s/it][I 2020-09-27 05:03:14,206] Trial 58 finished with value: 0.6880722255987741 and parameters: {'lambda_l1': 2.8108351646575752e-05, 'lambda_l2': 1.100109030564083e-08}. Best is trial 57 with value: 0.6880722255253509. regularization_factors, val_score: 0.688072: 80%|######## | 16/20 [00:24<00:06, 1.58s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012705 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486 [200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315 [300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232 [400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467 Early stopping, best iteration is: [334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072 regularization_factors, val_score: 0.688072: 85%|########5 | 17/20 [00:26<00:04, 1.57s/it][I 2020-09-27 05:03:15,742] Trial 59 finished with value: 0.6880722255024915 and parameters: {'lambda_l1': 3.5103404293664e-05, 'lambda_l2': 1.005446328322094e-08}. Best is trial 59 with value: 0.6880722255024915. regularization_factors, val_score: 0.688072: 85%|########5 | 17/20 [00:26<00:04, 1.57s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001474 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486 [200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315 [300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232 [400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467 Early stopping, best iteration is: [334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072 regularization_factors, val_score: 0.688072: 90%|######### | 18/20 [00:28<00:03, 1.58s/it][I 2020-09-27 05:03:17,361] Trial 60 finished with value: 0.6880722254979789 and parameters: {'lambda_l1': 3.541626511270535e-05, 'lambda_l2': 1.0717975785657199e-08}. Best is trial 60 with value: 0.6880722254979789. regularization_factors, val_score: 0.688072: 90%|######### | 18/20 [00:28<00:03, 1.58s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.010965 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486 [200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315 [300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232 [400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467 Early stopping, best iteration is: [334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072 regularization_factors, val_score: 0.688072: 95%|#########5| 19/20 [00:29<00:01, 1.58s/it][I 2020-09-27 05:03:18,921] Trial 61 finished with value: 0.6880722256490812 and parameters: {'lambda_l1': 2.4664030202905467e-05, 'lambda_l2': 1.0803279789819484e-08}. Best is trial 60 with value: 0.6880722254979789. regularization_factors, val_score: 0.688072: 95%|#########5| 19/20 [00:29<00:01, 1.58s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008152 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685985 valid's binary_logloss: 0.689486 [200] train's binary_logloss: 0.682684 valid's binary_logloss: 0.688315 [300] train's binary_logloss: 0.679961 valid's binary_logloss: 0.688232 [400] train's binary_logloss: 0.677324 valid's binary_logloss: 0.688467 Early stopping, best iteration is: [334] train's binary_logloss: 0.679066 valid's binary_logloss: 0.688072 regularization_factors, val_score: 0.688072: 100%|##########| 20/20 [00:31<00:00, 1.56s/it][I 2020-09-27 05:03:20,457] Trial 62 finished with value: 0.6880722251720996 and parameters: {'lambda_l1': 5.8781935567800015e-05, 'lambda_l2': 1.1695290416442605e-08}. Best is trial 62 with value: 0.6880722251720996. regularization_factors, val_score: 0.688072: 100%|##########| 20/20 [00:31<00:00, 1.56s/it] min_data_in_leaf, val_score: 0.688072: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.012689 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.686108 valid's binary_logloss: 0.689957 [200] train's binary_logloss: 0.683126 valid's binary_logloss: 0.688938 Early stopping, best iteration is: [199] train's binary_logloss: 0.683149 valid's binary_logloss: 0.688886 min_data_in_leaf, val_score: 0.688072: 20%|## | 1/5 [00:01<00:04, 1.14s/it][I 2020-09-27 05:03:21,608] Trial 63 finished with value: 0.6888860610141573 and parameters: {'min_child_samples': 100}. Best is trial 63 with value: 0.6888860610141573. min_data_in_leaf, val_score: 0.688072: 20%|## | 1/5 [00:01<00:04, 1.14s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009121 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685963 valid's binary_logloss: 0.689803 [200] train's binary_logloss: 0.682736 valid's binary_logloss: 0.688538 Early stopping, best iteration is: [198] train's binary_logloss: 0.682793 valid's binary_logloss: 0.688516 min_data_in_leaf, val_score: 0.688072: 40%|#### | 2/5 [00:02<00:03, 1.14s/it][I 2020-09-27 05:03:22,746] Trial 64 finished with value: 0.6885159620293074 and parameters: {'min_child_samples': 25}. Best is trial 64 with value: 0.6885159620293074. min_data_in_leaf, val_score: 0.688072: 40%|#### | 2/5 [00:02<00:03, 1.14s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008620 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68607 valid's binary_logloss: 0.689662 [200] train's binary_logloss: 0.682938 valid's binary_logloss: 0.688496 [300] train's binary_logloss: 0.68032 valid's binary_logloss: 0.688457 [400] train's binary_logloss: 0.677831 valid's binary_logloss: 0.688297 Early stopping, best iteration is: [330] train's binary_logloss: 0.679568 valid's binary_logloss: 0.6881 min_data_in_leaf, val_score: 0.688072: 60%|###### | 3/5 [00:03<00:02, 1.27s/it][I 2020-09-27 05:03:24,343] Trial 65 finished with value: 0.6880998602504527 and parameters: {'min_child_samples': 50}. Best is trial 65 with value: 0.6880998602504527. min_data_in_leaf, val_score: 0.688072: 60%|###### | 3/5 [00:03<00:02, 1.27s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001443 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685989 valid's binary_logloss: 0.689573 [200] train's binary_logloss: 0.682702 valid's binary_logloss: 0.688452 [300] train's binary_logloss: 0.679934 valid's binary_logloss: 0.68856 Early stopping, best iteration is: [272] train's binary_logloss: 0.680746 valid's binary_logloss: 0.688417 min_data_in_leaf, val_score: 0.688072: 80%|######## | 4/5 [00:05<00:01, 1.32s/it][I 2020-09-27 05:03:25,758] Trial 66 finished with value: 0.6884168700112442 and parameters: {'min_child_samples': 5}. Best is trial 65 with value: 0.6880998602504527. min_data_in_leaf, val_score: 0.688072: 80%|######## | 4/5 [00:05<00:01, 1.32s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008044 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.686007 valid's binary_logloss: 0.689624 [200] train's binary_logloss: 0.682678 valid's binary_logloss: 0.688328 [300] train's binary_logloss: 0.679893 valid's binary_logloss: 0.688247 [400] train's binary_logloss: 0.677249 valid's binary_logloss: 0.688678 Early stopping, best iteration is: [311] train's binary_logloss: 0.679598 valid's binary_logloss: 0.688186 min_data_in_leaf, val_score: 0.688072: 100%|##########| 5/5 [00:06<00:00, 1.38s/it][I 2020-09-27 05:03:27,278] Trial 67 finished with value: 0.6881857897062641 and parameters: {'min_child_samples': 10}. Best is trial 65 with value: 0.6880998602504527. min_data_in_leaf, val_score: 0.688072: 100%|##########| 5/5 [00:06<00:00, 1.36s/it] Fold : 9 [I 2020-09-27 05:03:27,416] A new study created in memory with name: no-name-e6a1e2b2-b26c-4a35-b26d-98523bba8f2b feature_fraction, val_score: inf: 0%| | 0/7 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001148 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664968 valid's binary_logloss: 0.689984 Early stopping, best iteration is: [38] train's binary_logloss: 0.679101 valid's binary_logloss: 0.689149 feature_fraction, val_score: 0.689149: 14%|#4 | 1/7 [00:01<00:06, 1.12s/it][I 2020-09-27 05:03:28,547] Trial 0 finished with value: 0.6891486028436887 and parameters: {'feature_fraction': 0.5}. Best is trial 0 with value: 0.6891486028436887. feature_fraction, val_score: 0.689149: 14%|#4 | 1/7 [00:01<00:06, 1.12s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.014784 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.662784 valid's binary_logloss: 0.690663 Early stopping, best iteration is: [40] train's binary_logloss: 0.677488 valid's binary_logloss: 0.689489 feature_fraction, val_score: 0.689149: 29%|##8 | 2/7 [00:02<00:05, 1.19s/it][I 2020-09-27 05:03:29,885] Trial 1 finished with value: 0.6894894237035651 and parameters: {'feature_fraction': 0.8999999999999999}. Best is trial 0 with value: 0.6891486028436887. feature_fraction, val_score: 0.689149: 29%|##8 | 2/7 [00:02<00:05, 1.19s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000849 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.66599 valid's binary_logloss: 0.69036 Early stopping, best iteration is: [64] train's binary_logloss: 0.67328 valid's binary_logloss: 0.689617 feature_fraction, val_score: 0.689149: 43%|####2 | 3/7 [00:03<00:04, 1.09s/it][I 2020-09-27 05:03:30,757] Trial 2 finished with value: 0.6896170710885305 and parameters: {'feature_fraction': 0.4}. Best is trial 0 with value: 0.6891486028436887. feature_fraction, val_score: 0.689149: 43%|####2 | 3/7 [00:03<00:04, 1.09s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.005632 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.66269 valid's binary_logloss: 0.690688 Early stopping, best iteration is: [38] train's binary_logloss: 0.67783 valid's binary_logloss: 0.689711 feature_fraction, val_score: 0.689149: 57%|#####7 | 4/7 [00:04<00:03, 1.03s/it][I 2020-09-27 05:03:31,639] Trial 3 finished with value: 0.6897107894464999 and parameters: {'feature_fraction': 1.0}. Best is trial 0 with value: 0.6891486028436887. feature_fraction, val_score: 0.689149: 57%|#####7 | 4/7 [00:04<00:03, 1.03s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013949 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.662985 valid's binary_logloss: 0.690369 Early stopping, best iteration is: [89] train's binary_logloss: 0.665582 valid's binary_logloss: 0.689673 feature_fraction, val_score: 0.689149: 71%|#######1 | 5/7 [00:05<00:02, 1.02s/it][I 2020-09-27 05:03:32,641] Trial 4 finished with value: 0.6896732000011675 and parameters: {'feature_fraction': 0.8}. Best is trial 0 with value: 0.6891486028436887. feature_fraction, val_score: 0.689149: 71%|#######1 | 5/7 [00:05<00:02, 1.02s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.007569 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.663886 valid's binary_logloss: 0.690174 Early stopping, best iteration is: [61] train's binary_logloss: 0.672601 valid's binary_logloss: 0.689421 feature_fraction, val_score: 0.689149: 86%|########5 | 6/7 [00:06<00:00, 1.04it/s][I 2020-09-27 05:03:33,477] Trial 5 finished with value: 0.6894210990140921 and parameters: {'feature_fraction': 0.7}. Best is trial 0 with value: 0.6891486028436887. feature_fraction, val_score: 0.689149: 86%|########5 | 6/7 [00:06<00:00, 1.04it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008826 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.664317 valid's binary_logloss: 0.690145 Early stopping, best iteration is: [52] train's binary_logloss: 0.675207 valid's binary_logloss: 0.689455 feature_fraction, val_score: 0.689149: 100%|##########| 7/7 [00:06<00:00, 1.11it/s][I 2020-09-27 05:03:34,239] Trial 6 finished with value: 0.6894553081219641 and parameters: {'feature_fraction': 0.6}. Best is trial 0 with value: 0.6891486028436887. feature_fraction, val_score: 0.689149: 100%|##########| 7/7 [00:06<00:00, 1.03it/s] num_leaves, val_score: 0.689149: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000950 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.564202 valid's binary_logloss: 0.694875 Early stopping, best iteration is: [19] train's binary_logloss: 0.658179 valid's binary_logloss: 0.69083 num_leaves, val_score: 0.689149: 5%|5 | 1/20 [00:01<00:23, 1.25s/it][I 2020-09-27 05:03:35,508] Trial 7 finished with value: 0.6908299824401397 and parameters: {'num_leaves': 204}. Best is trial 7 with value: 0.6908299824401397. num_leaves, val_score: 0.689149: 5%|5 | 1/20 [00:01<00:23, 1.25s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000951 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.564078 valid's binary_logloss: 0.696783 Early stopping, best iteration is: [13] train's binary_logloss: 0.668565 valid's binary_logloss: 0.690921 num_leaves, val_score: 0.689149: 10%|# | 2/20 [00:02<00:22, 1.23s/it][I 2020-09-27 05:03:36,684] Trial 8 finished with value: 0.6909207765265465 and parameters: {'num_leaves': 202}. Best is trial 7 with value: 0.6908299824401397. num_leaves, val_score: 0.689149: 10%|# | 2/20 [00:02<00:22, 1.23s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000865 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.625795 valid's binary_logloss: 0.693319 Early stopping, best iteration is: [35] train's binary_logloss: 0.663313 valid's binary_logloss: 0.690147 num_leaves, val_score: 0.689149: 15%|#5 | 3/20 [00:03<00:19, 1.13s/it][I 2020-09-27 05:03:37,591] Trial 9 finished with value: 0.6901469432883655 and parameters: {'num_leaves': 89}. Best is trial 9 with value: 0.6901469432883655. num_leaves, val_score: 0.689149: 15%|#5 | 3/20 [00:03<00:19, 1.13s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000928 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.690045 valid's binary_logloss: 0.690754 [200] train's binary_logloss: 0.688849 valid's binary_logloss: 0.690081 [300] train's binary_logloss: 0.688099 valid's binary_logloss: 0.6897 [400] train's binary_logloss: 0.687569 valid's binary_logloss: 0.689504 [500] train's binary_logloss: 0.687173 valid's binary_logloss: 0.689388 [600] train's binary_logloss: 0.686865 valid's binary_logloss: 0.689372 Early stopping, best iteration is: [537] train's binary_logloss: 0.687051 valid's binary_logloss: 0.689339 num_leaves, val_score: 0.689149: 20%|## | 4/20 [00:05<00:22, 1.41s/it][I 2020-09-27 05:03:39,636] Trial 10 finished with value: 0.6893392852593333 and parameters: {'num_leaves': 2}. Best is trial 10 with value: 0.6893392852593333. num_leaves, val_score: 0.689149: 20%|## | 4/20 [00:05<00:22, 1.41s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000928 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.686226 valid's binary_logloss: 0.689731 [200] train's binary_logloss: 0.683165 valid's binary_logloss: 0.689426 [300] train's binary_logloss: 0.680433 valid's binary_logloss: 0.689593 Early stopping, best iteration is: [221] train's binary_logloss: 0.682545 valid's binary_logloss: 0.689351 num_leaves, val_score: 0.689149: 25%|##5 | 5/20 [00:06<00:20, 1.34s/it][I 2020-09-27 05:03:40,838] Trial 11 finished with value: 0.6893514061615624 and parameters: {'num_leaves': 5}. Best is trial 10 with value: 0.6893392852593333. num_leaves, val_score: 0.689149: 25%|##5 | 5/20 [00:06<00:20, 1.34s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001034 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.677343 valid's binary_logloss: 0.689733 Early stopping, best iteration is: [73] train's binary_logloss: 0.680397 valid's binary_logloss: 0.689164 num_leaves, val_score: 0.689149: 30%|### | 6/20 [00:07<00:16, 1.18s/it][I 2020-09-27 05:03:41,643] Trial 12 finished with value: 0.6891638627451019 and parameters: {'num_leaves': 15}. Best is trial 12 with value: 0.6891638627451019. num_leaves, val_score: 0.689149: 30%|### | 6/20 [00:07<00:16, 1.18s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000978 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.625795 valid's binary_logloss: 0.693319 Early stopping, best iteration is: [35] train's binary_logloss: 0.663313 valid's binary_logloss: 0.690147 num_leaves, val_score: 0.689149: 35%|###5 | 7/20 [00:08<00:14, 1.12s/it][I 2020-09-27 05:03:42,610] Trial 13 finished with value: 0.6901469432883656 and parameters: {'num_leaves': 89}. Best is trial 12 with value: 0.6891638627451019. num_leaves, val_score: 0.689149: 35%|###5 | 7/20 [00:08<00:14, 1.12s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000900 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.645429 valid's binary_logloss: 0.691128 Early stopping, best iteration is: [37] train's binary_logloss: 0.670937 valid's binary_logloss: 0.689723 num_leaves, val_score: 0.689149: 40%|#### | 8/20 [00:09<00:12, 1.03s/it][I 2020-09-27 05:03:43,427] Trial 14 finished with value: 0.6897233852217974 and parameters: {'num_leaves': 58}. Best is trial 12 with value: 0.6891638627451019. num_leaves, val_score: 0.689149: 40%|#### | 8/20 [00:09<00:12, 1.03s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000972 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.53862 valid's binary_logloss: 0.696319 Early stopping, best iteration is: [19] train's binary_logloss: 0.651313 valid's binary_logloss: 0.690486 num_leaves, val_score: 0.689149: 45%|####5 | 9/20 [00:10<00:12, 1.13s/it][I 2020-09-27 05:03:44,806] Trial 15 finished with value: 0.6904857400051433 and parameters: {'num_leaves': 254}. Best is trial 12 with value: 0.6891638627451019. num_leaves, val_score: 0.689149: 45%|####5 | 9/20 [00:10<00:12, 1.13s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000921 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.58509 valid's binary_logloss: 0.697538 Early stopping, best iteration is: [18] train's binary_logloss: 0.665762 valid's binary_logloss: 0.690884 num_leaves, val_score: 0.689149: 50%|##### | 10/20 [00:11<00:11, 1.10s/it][I 2020-09-27 05:03:45,833] Trial 16 finished with value: 0.6908843676817171 and parameters: {'num_leaves': 160}. Best is trial 12 with value: 0.6891638627451019. num_leaves, val_score: 0.689149: 50%|##### | 10/20 [00:11<00:11, 1.10s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.008682 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.655506 valid's binary_logloss: 0.690438 Early stopping, best iteration is: [34] train's binary_logloss: 0.67629 valid's binary_logloss: 0.689323 num_leaves, val_score: 0.689149: 55%|#####5 | 11/20 [00:12<00:08, 1.03it/s][I 2020-09-27 05:03:46,509] Trial 17 finished with value: 0.6893229030060266 and parameters: {'num_leaves': 44}. Best is trial 12 with value: 0.6891638627451019. num_leaves, val_score: 0.689149: 55%|#####5 | 11/20 [00:12<00:08, 1.03it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000877 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.598808 valid's binary_logloss: 0.694092 Early stopping, best iteration is: [16] train's binary_logloss: 0.671679 valid's binary_logloss: 0.690451 num_leaves, val_score: 0.689149: 60%|###### | 12/20 [00:13<00:07, 1.02it/s][I 2020-09-27 05:03:47,497] Trial 18 finished with value: 0.6904513066474659 and parameters: {'num_leaves': 136}. Best is trial 12 with value: 0.6891638627451019. num_leaves, val_score: 0.689149: 60%|###### | 12/20 [00:13<00:07, 1.02it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000956 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.665649 valid's binary_logloss: 0.690234 Early stopping, best iteration is: [38] train's binary_logloss: 0.679544 valid's binary_logloss: 0.689555 num_leaves, val_score: 0.689149: 65%|######5 | 13/20 [00:13<00:06, 1.12it/s][I 2020-09-27 05:03:48,195] Trial 19 finished with value: 0.6895546123306635 and parameters: {'num_leaves': 30}. Best is trial 12 with value: 0.6891638627451019. num_leaves, val_score: 0.689149: 65%|######5 | 13/20 [00:13<00:06, 1.12it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000933 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.625105 valid's binary_logloss: 0.692852 Early stopping, best iteration is: [37] train's binary_logloss: 0.661371 valid's binary_logloss: 0.690145 num_leaves, val_score: 0.689149: 70%|####### | 14/20 [00:14<00:05, 1.09it/s][I 2020-09-27 05:03:49,159] Trial 20 finished with value: 0.6901453602236771 and parameters: {'num_leaves': 91}. Best is trial 12 with value: 0.6891638627451019. num_leaves, val_score: 0.689149: 70%|####### | 14/20 [00:14<00:05, 1.09it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000954 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.658357 valid's binary_logloss: 0.69009 Early stopping, best iteration is: [53] train's binary_logloss: 0.671442 valid's binary_logloss: 0.689611 num_leaves, val_score: 0.689149: 75%|#######5 | 15/20 [00:15<00:04, 1.12it/s][I 2020-09-27 05:03:49,991] Trial 21 finished with value: 0.6896105933356692 and parameters: {'num_leaves': 40}. Best is trial 12 with value: 0.6891638627451019. num_leaves, val_score: 0.689149: 75%|#######5 | 15/20 [00:15<00:04, 1.12it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000936 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685253 valid's binary_logloss: 0.689642 [200] train's binary_logloss: 0.681379 valid's binary_logloss: 0.689235 Early stopping, best iteration is: [185] train's binary_logloss: 0.681921 valid's binary_logloss: 0.689088 num_leaves, val_score: 0.689088: 80%|######## | 16/20 [00:16<00:03, 1.04it/s][I 2020-09-27 05:03:51,133] Trial 22 finished with value: 0.689088142033773 and parameters: {'num_leaves': 6}. Best is trial 22 with value: 0.689088142033773. num_leaves, val_score: 0.689088: 80%|######## | 16/20 [00:16<00:03, 1.04it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000953 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.687295 valid's binary_logloss: 0.68994 [200] train's binary_logloss: 0.684807 valid's binary_logloss: 0.689595 [300] train's binary_logloss: 0.682867 valid's binary_logloss: 0.689608 Early stopping, best iteration is: [256] train's binary_logloss: 0.683708 valid's binary_logloss: 0.689506 num_leaves, val_score: 0.689088: 85%|########5 | 17/20 [00:18<00:03, 1.06s/it][I 2020-09-27 05:03:52,425] Trial 23 finished with value: 0.6895057052158261 and parameters: {'num_leaves': 4}. Best is trial 22 with value: 0.689088142033773. num_leaves, val_score: 0.689088: 85%|########5 | 17/20 [00:18<00:03, 1.06s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011377 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.636575 valid's binary_logloss: 0.690665 Early stopping, best iteration is: [36] train's binary_logloss: 0.667345 valid's binary_logloss: 0.689716 num_leaves, val_score: 0.689088: 90%|######### | 18/20 [00:18<00:01, 1.01it/s][I 2020-09-27 05:03:53,255] Trial 24 finished with value: 0.6897160545946952 and parameters: {'num_leaves': 72}. Best is trial 22 with value: 0.689088142033773. num_leaves, val_score: 0.689088: 90%|######### | 18/20 [00:19<00:01, 1.01it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000933 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.667234 valid's binary_logloss: 0.689986 Early stopping, best iteration is: [74] train's binary_logloss: 0.672163 valid's binary_logloss: 0.689426 num_leaves, val_score: 0.689088: 95%|#########5| 19/20 [00:19<00:00, 1.05it/s][I 2020-09-27 05:03:54,104] Trial 25 finished with value: 0.6894258959836473 and parameters: {'num_leaves': 28}. Best is trial 22 with value: 0.689088142033773. num_leaves, val_score: 0.689088: 95%|#########5| 19/20 [00:19<00:00, 1.05it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000884 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.609273 valid's binary_logloss: 0.693331 Early stopping, best iteration is: [34] train's binary_logloss: 0.656448 valid's binary_logloss: 0.689954 num_leaves, val_score: 0.689088: 100%|##########| 20/20 [00:20<00:00, 1.02it/s][I 2020-09-27 05:03:55,150] Trial 26 finished with value: 0.6899541998715152 and parameters: {'num_leaves': 117}. Best is trial 22 with value: 0.689088142033773. num_leaves, val_score: 0.689088: 100%|##########| 20/20 [00:20<00:00, 1.04s/it] bagging, val_score: 0.689088: 0%| | 0/10 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000959 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685151 valid's binary_logloss: 0.689666 [200] train's binary_logloss: 0.681327 valid's binary_logloss: 0.689749 Early stopping, best iteration is: [143] train's binary_logloss: 0.68343 valid's binary_logloss: 0.689482 bagging, val_score: 0.689088: 10%|# | 1/10 [00:01<00:09, 1.05s/it][I 2020-09-27 05:03:56,212] Trial 27 finished with value: 0.6894817163491044 and parameters: {'bagging_fraction': 0.9223383259961062, 'bagging_freq': 3}. Best is trial 27 with value: 0.6894817163491044. bagging, val_score: 0.689088: 10%|# | 1/10 [00:01<00:09, 1.05s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000943 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685286 valid's binary_logloss: 0.689382 [200] train's binary_logloss: 0.68151 valid's binary_logloss: 0.689555 Early stopping, best iteration is: [173] train's binary_logloss: 0.682462 valid's binary_logloss: 0.688945 bagging, val_score: 0.688945: 20%|## | 2/10 [00:02<00:08, 1.04s/it][I 2020-09-27 05:03:57,220] Trial 28 finished with value: 0.6889445920177824 and parameters: {'bagging_fraction': 0.40058272142371876, 'bagging_freq': 7}. Best is trial 28 with value: 0.6889445920177824. bagging, val_score: 0.688945: 20%|## | 2/10 [00:02<00:08, 1.04s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000924 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685224 valid's binary_logloss: 0.689275 [200] train's binary_logloss: 0.681491 valid's binary_logloss: 0.689706 Early stopping, best iteration is: [168] train's binary_logloss: 0.682543 valid's binary_logloss: 0.688999 bagging, val_score: 0.688945: 30%|### | 3/10 [00:03<00:07, 1.03s/it][I 2020-09-27 05:03:58,234] Trial 29 finished with value: 0.6889992638135699 and parameters: {'bagging_fraction': 0.4181787978927057, 'bagging_freq': 7}. Best is trial 28 with value: 0.6889445920177824. bagging, val_score: 0.688945: 30%|### | 3/10 [00:03<00:07, 1.03s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000991 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685284 valid's binary_logloss: 0.68908 [200] train's binary_logloss: 0.681564 valid's binary_logloss: 0.689249 Early stopping, best iteration is: [172] train's binary_logloss: 0.682574 valid's binary_logloss: 0.688831 bagging, val_score: 0.688831: 40%|#### | 4/10 [00:04<00:06, 1.02s/it][I 2020-09-27 05:03:59,225] Trial 30 finished with value: 0.688830953603927 and parameters: {'bagging_fraction': 0.40441196806803215, 'bagging_freq': 7}. Best is trial 30 with value: 0.688830953603927. bagging, val_score: 0.688831: 40%|#### | 4/10 [00:04<00:06, 1.02s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001020 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685401 valid's binary_logloss: 0.689413 [200] train's binary_logloss: 0.68163 valid's binary_logloss: 0.6894 Early stopping, best iteration is: [172] train's binary_logloss: 0.682614 valid's binary_logloss: 0.688975 bagging, val_score: 0.688831: 50%|##### | 5/10 [00:05<00:05, 1.01s/it][I 2020-09-27 05:04:00,231] Trial 31 finished with value: 0.688974864123655 and parameters: {'bagging_fraction': 0.4053863270403766, 'bagging_freq': 7}. Best is trial 30 with value: 0.688830953603927. bagging, val_score: 0.688831: 50%|##### | 5/10 [00:05<00:05, 1.01s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000972 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685259 valid's binary_logloss: 0.68914 Early stopping, best iteration is: [95] train's binary_logloss: 0.685454 valid's binary_logloss: 0.689029 bagging, val_score: 0.688831: 60%|###### | 6/10 [00:05<00:03, 1.07it/s][I 2020-09-27 05:04:00,968] Trial 32 finished with value: 0.6890288131482014 and parameters: {'bagging_fraction': 0.4233037496739684, 'bagging_freq': 7}. Best is trial 30 with value: 0.688830953603927. bagging, val_score: 0.688831: 60%|###### | 6/10 [00:05<00:03, 1.07it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000937 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685245 valid's binary_logloss: 0.689157 Early stopping, best iteration is: [87] train's binary_logloss: 0.685847 valid's binary_logloss: 0.688973 bagging, val_score: 0.688831: 70%|####### | 7/10 [00:06<00:02, 1.16it/s][I 2020-09-27 05:04:01,669] Trial 33 finished with value: 0.6889732179547475 and parameters: {'bagging_fraction': 0.4054479427416367, 'bagging_freq': 7}. Best is trial 30 with value: 0.688830953603927. bagging, val_score: 0.688831: 70%|####### | 7/10 [00:06<00:02, 1.16it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001013 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685369 valid's binary_logloss: 0.688941 [200] train's binary_logloss: 0.681625 valid's binary_logloss: 0.689259 Early stopping, best iteration is: [174] train's binary_logloss: 0.682563 valid's binary_logloss: 0.688665 bagging, val_score: 0.688665: 80%|######## | 8/10 [00:07<00:01, 1.12it/s][I 2020-09-27 05:04:02,643] Trial 34 finished with value: 0.6886646750149633 and parameters: {'bagging_fraction': 0.4052909923895454, 'bagging_freq': 7}. Best is trial 34 with value: 0.6886646750149633. bagging, val_score: 0.688665: 80%|######## | 8/10 [00:07<00:01, 1.12it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001120 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68534 valid's binary_logloss: 0.68936 [200] train's binary_logloss: 0.681589 valid's binary_logloss: 0.689277 Early stopping, best iteration is: [160] train's binary_logloss: 0.682993 valid's binary_logloss: 0.689007 bagging, val_score: 0.688665: 90%|######### | 9/10 [00:08<00:00, 1.09it/s][I 2020-09-27 05:04:03,602] Trial 35 finished with value: 0.6890068245461886 and parameters: {'bagging_fraction': 0.40351986086829467, 'bagging_freq': 7}. Best is trial 34 with value: 0.6886646750149633. bagging, val_score: 0.688665: 90%|######### | 9/10 [00:08<00:00, 1.09it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000983 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68515 valid's binary_logloss: 0.68937 [200] train's binary_logloss: 0.681294 valid's binary_logloss: 0.689514 Early stopping, best iteration is: [113] train's binary_logloss: 0.684606 valid's binary_logloss: 0.68927 bagging, val_score: 0.688665: 100%|##########| 10/10 [00:09<00:00, 1.11it/s][I 2020-09-27 05:04:04,460] Trial 36 finished with value: 0.6892703778547441 and parameters: {'bagging_fraction': 0.5684326008067807, 'bagging_freq': 5}. Best is trial 34 with value: 0.6886646750149633. bagging, val_score: 0.688665: 100%|##########| 10/10 [00:09<00:00, 1.08it/s] feature_fraction_stage2, val_score: 0.688665: 0%| | 0/6 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000961 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685369 valid's binary_logloss: 0.688941 [200] train's binary_logloss: 0.681625 valid's binary_logloss: 0.689259 Early stopping, best iteration is: [174] train's binary_logloss: 0.682563 valid's binary_logloss: 0.688665 feature_fraction_stage2, val_score: 0.688665: 17%|#6 | 1/6 [00:01<00:05, 1.01s/it][I 2020-09-27 05:04:05,490] Trial 37 finished with value: 0.6886646750149633 and parameters: {'feature_fraction': 0.516}. Best is trial 37 with value: 0.6886646750149633. feature_fraction_stage2, val_score: 0.688665: 17%|#6 | 1/6 [00:01<00:05, 1.01s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000867 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685291 valid's binary_logloss: 0.689179 Early stopping, best iteration is: [86] train's binary_logloss: 0.685898 valid's binary_logloss: 0.688923 feature_fraction_stage2, val_score: 0.688665: 33%|###3 | 2/6 [00:01<00:03, 1.09it/s][I 2020-09-27 05:04:06,195] Trial 38 finished with value: 0.6889230860549193 and parameters: {'feature_fraction': 0.45199999999999996}. Best is trial 37 with value: 0.6886646750149633. feature_fraction_stage2, val_score: 0.688665: 33%|###3 | 2/6 [00:01<00:03, 1.09it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001184 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685287 valid's binary_logloss: 0.689043 [200] train's binary_logloss: 0.681498 valid's binary_logloss: 0.689164 Early stopping, best iteration is: [168] train's binary_logloss: 0.682602 valid's binary_logloss: 0.688738 feature_fraction_stage2, val_score: 0.688665: 50%|##### | 3/6 [00:02<00:02, 1.07it/s][I 2020-09-27 05:04:07,161] Trial 39 finished with value: 0.6887381070029943 and parameters: {'feature_fraction': 0.5479999999999999}. Best is trial 37 with value: 0.6886646750149633. feature_fraction_stage2, val_score: 0.688665: 50%|##### | 3/6 [00:02<00:02, 1.07it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000856 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685333 valid's binary_logloss: 0.689331 [200] train's binary_logloss: 0.681639 valid's binary_logloss: 0.689698 Early stopping, best iteration is: [166] train's binary_logloss: 0.682754 valid's binary_logloss: 0.689155 feature_fraction_stage2, val_score: 0.688665: 67%|######6 | 4/6 [00:03<00:01, 1.07it/s][I 2020-09-27 05:04:08,093] Trial 40 finished with value: 0.6891552510228472 and parameters: {'feature_fraction': 0.42}. Best is trial 37 with value: 0.6886646750149633. feature_fraction_stage2, val_score: 0.688665: 67%|######6 | 4/6 [00:03<00:01, 1.07it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001100 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685198 valid's binary_logloss: 0.689125 [200] train's binary_logloss: 0.681316 valid's binary_logloss: 0.689341 Early stopping, best iteration is: [174] train's binary_logloss: 0.682274 valid's binary_logloss: 0.688933 feature_fraction_stage2, val_score: 0.688665: 83%|########3 | 5/6 [00:04<00:00, 1.04it/s][I 2020-09-27 05:04:09,134] Trial 41 finished with value: 0.6889333127717522 and parameters: {'feature_fraction': 0.58}. Best is trial 37 with value: 0.6886646750149633. feature_fraction_stage2, val_score: 0.688665: 83%|########3 | 5/6 [00:04<00:00, 1.04it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011122 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685369 valid's binary_logloss: 0.688941 [200] train's binary_logloss: 0.681625 valid's binary_logloss: 0.689259 Early stopping, best iteration is: [174] train's binary_logloss: 0.682563 valid's binary_logloss: 0.688665 feature_fraction_stage2, val_score: 0.688665: 100%|##########| 6/6 [00:05<00:00, 1.03it/s][I 2020-09-27 05:04:10,124] Trial 42 finished with value: 0.6886646750149633 and parameters: {'feature_fraction': 0.484}. Best is trial 37 with value: 0.6886646750149633. feature_fraction_stage2, val_score: 0.688665: 100%|##########| 6/6 [00:05<00:00, 1.06it/s] regularization_factors, val_score: 0.688665: 0%| | 0/20 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000996 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685322 valid's binary_logloss: 0.689114 [200] train's binary_logloss: 0.681508 valid's binary_logloss: 0.689186 Early stopping, best iteration is: [175] train's binary_logloss: 0.682376 valid's binary_logloss: 0.688631 regularization_factors, val_score: 0.688631: 5%|5 | 1/20 [00:01<00:20, 1.08s/it][I 2020-09-27 05:04:11,228] Trial 43 finished with value: 0.6886306062996863 and parameters: {'lambda_l1': 0.05593044522293832, 'lambda_l2': 0.003082206943612091}. Best is trial 43 with value: 0.6886306062996863. regularization_factors, val_score: 0.688631: 5%|5 | 1/20 [00:01<00:20, 1.08s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000926 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685295 valid's binary_logloss: 0.689098 Early stopping, best iteration is: [86] train's binary_logloss: 0.68591 valid's binary_logloss: 0.688875 regularization_factors, val_score: 0.688631: 10%|# | 2/20 [00:01<00:17, 1.01it/s][I 2020-09-27 05:04:11,991] Trial 44 finished with value: 0.6888752025477913 and parameters: {'lambda_l1': 0.1570663292442244, 'lambda_l2': 0.0010606198676237839}. Best is trial 43 with value: 0.6886306062996863. regularization_factors, val_score: 0.688631: 10%|# | 2/20 [00:01<00:17, 1.01it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000960 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.68561 valid's binary_logloss: 0.689202 Early stopping, best iteration is: [89] train's binary_logloss: 0.686012 valid's binary_logloss: 0.688996 regularization_factors, val_score: 0.688631: 15%|#5 | 3/20 [00:02<00:15, 1.10it/s][I 2020-09-27 05:04:12,715] Trial 45 finished with value: 0.6889959749240171 and parameters: {'lambda_l1': 1.4927818545611396e-06, 'lambda_l2': 7.613862377908525}. Best is trial 43 with value: 0.6886306062996863. regularization_factors, val_score: 0.688631: 15%|#5 | 3/20 [00:02<00:15, 1.10it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000944 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685965 valid's binary_logloss: 0.689511 [200] train's binary_logloss: 0.682794 valid's binary_logloss: 0.689793 Early stopping, best iteration is: [174] train's binary_logloss: 0.683538 valid's binary_logloss: 0.689222 regularization_factors, val_score: 0.688631: 20%|## | 4/20 [00:03<00:15, 1.05it/s][I 2020-09-27 05:04:13,760] Trial 46 finished with value: 0.6892218212122403 and parameters: {'lambda_l1': 5.48610084980033, 'lambda_l2': 0.00024090694580484778}. Best is trial 43 with value: 0.6886306062996863. regularization_factors, val_score: 0.688631: 20%|## | 4/20 [00:03<00:15, 1.05it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.004926 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685369 valid's binary_logloss: 0.688941 [200] train's binary_logloss: 0.681626 valid's binary_logloss: 0.689259 Early stopping, best iteration is: [174] train's binary_logloss: 0.682563 valid's binary_logloss: 0.688665 regularization_factors, val_score: 0.688631: 25%|##5 | 5/20 [00:04<00:14, 1.02it/s][I 2020-09-27 05:04:14,825] Trial 47 finished with value: 0.6886646716779263 and parameters: {'lambda_l1': 0.0011713348102023722, 'lambda_l2': 1.4326223733321923e-08}. Best is trial 43 with value: 0.6886306062996863. regularization_factors, val_score: 0.688631: 25%|##5 | 5/20 [00:04<00:14, 1.02it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001038 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685369 valid's binary_logloss: 0.688941 [200] train's binary_logloss: 0.681626 valid's binary_logloss: 0.689259 Early stopping, best iteration is: [174] train's binary_logloss: 0.682563 valid's binary_logloss: 0.688665 regularization_factors, val_score: 0.688631: 30%|### | 6/20 [00:05<00:14, 1.07s/it][I 2020-09-27 05:04:16,093] Trial 48 finished with value: 0.6886646717918603 and parameters: {'lambda_l1': 0.0011323335612472105, 'lambda_l2': 1.2449590491320676e-08}. Best is trial 43 with value: 0.6886306062996863. regularization_factors, val_score: 0.688631: 30%|### | 6/20 [00:05<00:14, 1.07s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.009231 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685369 valid's binary_logloss: 0.688941 [200] train's binary_logloss: 0.681626 valid's binary_logloss: 0.689259 Early stopping, best iteration is: [174] train's binary_logloss: 0.682563 valid's binary_logloss: 0.688665 regularization_factors, val_score: 0.688631: 35%|###5 | 7/20 [00:06<00:13, 1.05s/it][I 2020-09-27 05:04:17,084] Trial 49 finished with value: 0.6886646722615538 and parameters: {'lambda_l1': 0.0009667778650190732, 'lambda_l2': 1.4964282342181056e-08}. Best is trial 43 with value: 0.6886306062996863. regularization_factors, val_score: 0.688631: 35%|###5 | 7/20 [00:06<00:13, 1.05s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001006 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685369 valid's binary_logloss: 0.688941 [200] train's binary_logloss: 0.681626 valid's binary_logloss: 0.689259 Early stopping, best iteration is: [174] train's binary_logloss: 0.682563 valid's binary_logloss: 0.688665 regularization_factors, val_score: 0.688631: 40%|#### | 8/20 [00:07<00:12, 1.04s/it][I 2020-09-27 05:04:18,110] Trial 50 finished with value: 0.6886646726282439 and parameters: {'lambda_l1': 0.000834005476706058, 'lambda_l2': 1.1339953988984425e-08}. Best is trial 43 with value: 0.6886306062996863. regularization_factors, val_score: 0.688631: 40%|#### | 8/20 [00:07<00:12, 1.04s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001002 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685369 valid's binary_logloss: 0.688941 [200] train's binary_logloss: 0.681626 valid's binary_logloss: 0.689259 Early stopping, best iteration is: [174] train's binary_logloss: 0.682563 valid's binary_logloss: 0.688665 regularization_factors, val_score: 0.688631: 45%|####5 | 9/20 [00:09<00:11, 1.04s/it][I 2020-09-27 05:04:19,146] Trial 51 finished with value: 0.6886646724473121 and parameters: {'lambda_l1': 0.0008969854861021092, 'lambda_l2': 1.1117633466286682e-08}. Best is trial 43 with value: 0.6886306062996863. regularization_factors, val_score: 0.688631: 45%|####5 | 9/20 [00:09<00:11, 1.04s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.013191 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685369 valid's binary_logloss: 0.688941 [200] train's binary_logloss: 0.681626 valid's binary_logloss: 0.689259 Early stopping, best iteration is: [174] train's binary_logloss: 0.682563 valid's binary_logloss: 0.688665 regularization_factors, val_score: 0.688631: 50%|##### | 10/20 [00:09<00:10, 1.01s/it][I 2020-09-27 05:04:20,098] Trial 52 finished with value: 0.6886646724367362 and parameters: {'lambda_l1': 0.0009016551190739346, 'lambda_l2': 1.0925317395658214e-08}. Best is trial 43 with value: 0.6886306062996863. regularization_factors, val_score: 0.688631: 50%|##### | 10/20 [00:09<00:10, 1.01s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000952 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685369 valid's binary_logloss: 0.688941 [200] train's binary_logloss: 0.681626 valid's binary_logloss: 0.689259 Early stopping, best iteration is: [174] train's binary_logloss: 0.682563 valid's binary_logloss: 0.688665 regularization_factors, val_score: 0.688631: 55%|#####5 | 11/20 [00:10<00:09, 1.01s/it][I 2020-09-27 05:04:21,108] Trial 53 finished with value: 0.6886646727381388 and parameters: {'lambda_l1': 0.0007965344637915165, 'lambda_l2': 1.3659640884572439e-08}. Best is trial 43 with value: 0.6886306062996863. regularization_factors, val_score: 0.688631: 55%|#####5 | 11/20 [00:10<00:09, 1.01s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.011699 seconds. You can set `force_col_wise=true` to remove the overhead. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685369 valid's binary_logloss: 0.688941 [200] train's binary_logloss: 0.681626 valid's binary_logloss: 0.689259 Early stopping, best iteration is: [174] train's binary_logloss: 0.682563 valid's binary_logloss: 0.688665 regularization_factors, val_score: 0.688631: 60%|###### | 12/20 [00:11<00:07, 1.01it/s][I 2020-09-27 05:04:22,056] Trial 54 finished with value: 0.6886646716630922 and parameters: {'lambda_l1': 0.0011776384915221255, 'lambda_l2': 1.0468618620438098e-08}. Best is trial 43 with value: 0.6886306062996863. regularization_factors, val_score: 0.688631: 60%|###### | 12/20 [00:11<00:07, 1.01it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000972 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685369 valid's binary_logloss: 0.688941 [200] train's binary_logloss: 0.681589 valid's binary_logloss: 0.689058 Early stopping, best iteration is: [174] train's binary_logloss: 0.68257 valid's binary_logloss: 0.688662 regularization_factors, val_score: 0.688631: 65%|######5 | 13/20 [00:12<00:06, 1.00it/s][I 2020-09-27 05:04:23,068] Trial 55 finished with value: 0.6886618433996927 and parameters: {'lambda_l1': 0.00344335006832809, 'lambda_l2': 1.0707613854300203e-08}. Best is trial 43 with value: 0.6886306062996863. regularization_factors, val_score: 0.688631: 65%|######5 | 13/20 [00:12<00:06, 1.00it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000963 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685283 valid's binary_logloss: 0.689021 [200] train's binary_logloss: 0.68165 valid's binary_logloss: 0.688997 Early stopping, best iteration is: [172] train's binary_logloss: 0.682609 valid's binary_logloss: 0.688401 regularization_factors, val_score: 0.688401: 70%|####### | 14/20 [00:13<00:05, 1.00it/s][I 2020-09-27 05:04:24,069] Trial 56 finished with value: 0.6884008811372169 and parameters: {'lambda_l1': 0.013564067517829166, 'lambda_l2': 0.0028990672498837708}. Best is trial 56 with value: 0.6884008811372169. regularization_factors, val_score: 0.688401: 70%|####### | 14/20 [00:13<00:05, 1.00it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001378 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685305 valid's binary_logloss: 0.689126 [200] train's binary_logloss: 0.681644 valid's binary_logloss: 0.689459 Early stopping, best iteration is: [187] train's binary_logloss: 0.68211 valid's binary_logloss: 0.688878 regularization_factors, val_score: 0.688401: 75%|#######5 | 15/20 [00:14<00:05, 1.02s/it][I 2020-09-27 05:04:25,134] Trial 57 finished with value: 0.6888781933882588 and parameters: {'lambda_l1': 0.06847066607771893, 'lambda_l2': 0.005100079808171161}. Best is trial 56 with value: 0.6884008811372169. regularization_factors, val_score: 0.688401: 75%|#######5 | 15/20 [00:14<00:05, 1.02s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000979 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685375 valid's binary_logloss: 0.689304 [200] train's binary_logloss: 0.681532 valid's binary_logloss: 0.689352 Early stopping, best iteration is: [173] train's binary_logloss: 0.682496 valid's binary_logloss: 0.688928 regularization_factors, val_score: 0.688401: 80%|######## | 16/20 [00:15<00:04, 1.01s/it][I 2020-09-27 05:04:26,139] Trial 58 finished with value: 0.6889282775916249 and parameters: {'lambda_l1': 0.04206907981702274, 'lambda_l2': 0.008049472807870911}. Best is trial 56 with value: 0.6884008811372169. regularization_factors, val_score: 0.688401: 80%|######## | 16/20 [00:15<00:04, 1.01s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001050 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685283 valid's binary_logloss: 0.689021 [200] train's binary_logloss: 0.68165 valid's binary_logloss: 0.688997 Early stopping, best iteration is: [172] train's binary_logloss: 0.682609 valid's binary_logloss: 0.688401 regularization_factors, val_score: 0.688401: 85%|########5 | 17/20 [00:19<00:05, 1.68s/it][I 2020-09-27 05:04:29,386] Trial 59 finished with value: 0.6884009077923067 and parameters: {'lambda_l1': 0.01818292826675875, 'lambda_l2': 4.793765360259705e-07}. Best is trial 56 with value: 0.6884008811372169. regularization_factors, val_score: 0.688401: 85%|########5 | 17/20 [00:19<00:05, 1.68s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001800 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685284 valid's binary_logloss: 0.689021 [200] train's binary_logloss: 0.681652 valid's binary_logloss: 0.688997 Early stopping, best iteration is: [172] train's binary_logloss: 0.68261 valid's binary_logloss: 0.688401 regularization_factors, val_score: 0.688401: 90%|######### | 18/20 [00:20<00:03, 1.51s/it][I 2020-09-27 05:04:30,486] Trial 60 finished with value: 0.6884009815019274 and parameters: {'lambda_l1': 0.022406843035860227, 'lambda_l2': 2.0488930606948696e-06}. Best is trial 56 with value: 0.6884008811372169. regularization_factors, val_score: 0.688401: 90%|######### | 18/20 [00:20<00:03, 1.51s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001595 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685284 valid's binary_logloss: 0.689021 [200] train's binary_logloss: 0.681652 valid's binary_logloss: 0.688997 Early stopping, best iteration is: [172] train's binary_logloss: 0.682611 valid's binary_logloss: 0.688401 regularization_factors, val_score: 0.688401: 95%|#########5| 19/20 [00:21<00:01, 1.39s/it][I 2020-09-27 05:04:31,591] Trial 61 finished with value: 0.6884010228456562 and parameters: {'lambda_l1': 0.024768099089603052, 'lambda_l2': 5.962389419495749e-06}. Best is trial 56 with value: 0.6884008811372169. regularization_factors, val_score: 0.688401: 95%|#########5| 19/20 [00:21<00:01, 1.39s/it][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001134 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685378 valid's binary_logloss: 0.688831 [200] train's binary_logloss: 0.681663 valid's binary_logloss: 0.688838 Early stopping, best iteration is: [182] train's binary_logloss: 0.682288 valid's binary_logloss: 0.688653 regularization_factors, val_score: 0.688401: 100%|##########| 20/20 [00:22<00:00, 1.32s/it][I 2020-09-27 05:04:32,742] Trial 62 finished with value: 0.6886532224197758 and parameters: {'lambda_l1': 0.03746482418394038, 'lambda_l2': 1.895168169019993e-06}. Best is trial 56 with value: 0.6884008811372169. regularization_factors, val_score: 0.688401: 100%|##########| 20/20 [00:22<00:00, 1.13s/it] min_data_in_leaf, val_score: 0.688401: 0%| | 0/5 [00:00<?, ?it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000979 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685537 valid's binary_logloss: 0.689132 Early stopping, best iteration is: [95] train's binary_logloss: 0.68572 valid's binary_logloss: 0.689056 min_data_in_leaf, val_score: 0.688401: 20%|## | 1/5 [00:00<00:03, 1.30it/s][I 2020-09-27 05:04:33,525] Trial 63 finished with value: 0.6890556326625765 and parameters: {'min_child_samples': 100}. Best is trial 63 with value: 0.6890556326625765. min_data_in_leaf, val_score: 0.688401: 20%|## | 1/5 [00:00<00:03, 1.30it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000984 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685507 valid's binary_logloss: 0.688911 Early stopping, best iteration is: [86] train's binary_logloss: 0.68607 valid's binary_logloss: 0.688678 min_data_in_leaf, val_score: 0.688401: 40%|#### | 2/5 [00:01<00:02, 1.32it/s][I 2020-09-27 05:04:34,254] Trial 64 finished with value: 0.68867811407117 and parameters: {'min_child_samples': 50}. Best is trial 64 with value: 0.68867811407117. min_data_in_leaf, val_score: 0.688401: 40%|#### | 2/5 [00:01<00:02, 1.32it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001032 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685329 valid's binary_logloss: 0.688996 [200] train's binary_logloss: 0.681584 valid's binary_logloss: 0.689202 Early stopping, best iteration is: [168] train's binary_logloss: 0.682673 valid's binary_logloss: 0.688789 min_data_in_leaf, val_score: 0.688401: 60%|###### | 3/5 [00:02<00:01, 1.20it/s][I 2020-09-27 05:04:35,256] Trial 65 finished with value: 0.6887885221160116 and parameters: {'min_child_samples': 25}. Best is trial 64 with value: 0.68867811407117. min_data_in_leaf, val_score: 0.688401: 60%|###### | 3/5 [00:02<00:01, 1.20it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.001002 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685286 valid's binary_logloss: 0.689236 Early stopping, best iteration is: [87] train's binary_logloss: 0.685815 valid's binary_logloss: 0.68908 min_data_in_leaf, val_score: 0.688401: 80%|######## | 4/5 [00:03<00:00, 1.27it/s][I 2020-09-27 05:04:35,953] Trial 66 finished with value: 0.6890800984264625 and parameters: {'min_child_samples': 10}. Best is trial 64 with value: 0.68867811407117. min_data_in_leaf, val_score: 0.688401: 80%|######## | 4/5 [00:03<00:00, 1.27it/s][LightGBM] [Info] Number of positive: 46363, number of negative: 46663 [LightGBM] [Warning] Auto-choosing row-wise multi-threading, the overhead of testing was 0.000956 seconds. You can set `force_row_wise=true` to remove the overhead. And if memory is not enough, you can set `force_col_wise=true`. [LightGBM] [Info] Total Bins 4689 [LightGBM] [Info] Number of data points in the train set: 93026, number of used features: 26 [LightGBM] [Info] [binary:BoostFromScore]: pavg=0.498388 -> initscore=-0.006450 [LightGBM] [Info] Start training from score -0.006450 Training until validation scores don't improve for 100 rounds [100] train's binary_logloss: 0.685281 valid's binary_logloss: 0.688938 Early stopping, best iteration is: [96] train's binary_logloss: 0.685434 valid's binary_logloss: 0.688858 min_data_in_leaf, val_score: 0.688401: 100%|##########| 5/5 [00:03<00:00, 1.29it/s][I 2020-09-27 05:04:36,693] Trial 67 finished with value: 0.6888576677735015 and parameters: {'min_child_samples': 5}. Best is trial 64 with value: 0.68867811407117. min_data_in_leaf, val_score: 0.688401: 100%|##########| 5/5 [00:03<00:00, 1.27it/s] ################################ CV_score:0.5369671988207919 --------------------------------- total CV_score:0.5760159134509492
# 各モードの予測結果を結合する
test_all = pd.concat([test_regular_id, test_gachi_id])
# id列でソートする
test_all = test_all.sort_values('id')
# 提出用ファイルを作成する
test_all.to_csv("submission.csv", index=False)
sylk
4-1の欠損値の件は、A1-rankなどのrank系かも。
たびたびでてますが、rankは「ガチマッチ」のウデマエが割り当てられており、ナワバリバトルには存在しませんので、全部nullが入ってます。 nullしかない特徴量をターゲットエンコーディングできるのかわからないですが、そうかも。
cha_kabu
スプラトゥーンをプレイしたことが無いのでアイデア助かります。ありがとうございます。
欠損について、恐らくですがmodeで分けてしまうとブキ種によってはデータ数が十分でなく、さらにそれをFold分割した際にseedによっては該当データが無くなってしまい、Target Encodingに必要な平均値が出せていないのではないでしょうか。
以下のブログ記事が参考になったのでご参考ください。 https://blog.amedama.jp/entry/target-mean-encoding-types
しかしその問題が起きるのは使用率が低いブキ種だけですし、プレイヤー毎ではなくチーム毎にEncordingするなど対処法はありそうです(まだ試していないので効果があるかは分かりません…)。
t_mochizuki0
欠損値の問題は保留とし、最終的な提出用データを作るところまで作成しました。
結果を提出したところ、全データをまとめて学習した時よりPublicスコアが向上しました。
上に表示されているコードを更新しました。
実験なので力技で実装しましたが、参考になれば幸いです。
欠損値は、ゆっくり確認します…。
t_mochizuki0
データ(と学習)の分割ですが、ガチエリア〜ガチヤグラの4種類に分けると細かすぎて学習データが減るのか、CVが低いモードがあります(ガチヤグラ)。
このためレギュラーマッチ(=ナワバリバトル)とガチマッチ(ガチエリア〜ガチヤグラ)の2種類に分けてみました。
Publicスコアはこちらの方が若干高いので、これを最終版とします。ありがとうございました。
(欠損値の問題は、ゆっくり確認します)
yktsnd
勉強させていただいております。
気がついた点が一点あるのでコメントさせてください。