LightGBM与评分卡

调参策略

最大化 off_ks + 0.8(off_ks-train_ks)

import pandas as pd
from sklearn.metrics import roc_auc_score,roc_curve,auc
from sklearn.model_selection import train_test_split
from sklearn import metrics
from sklearn.linear_model import LogisticRegression
import numpy as np
import random
import math
import time
import lightgbm as lgb
data = pd.read_csv('Acard.txt')
data.head()
obs_mth bad_ind uid td_score jxl_score mj_score rh_score zzc_score zcx_score person_info finance_info credit_info act_info
0 2018-10-31 0.0 A10000005 0.675349 0.144072 0.186899 0.483640 0.928328 0.369644 -0.322581 0.023810 0.00 0.217949
1 2018-07-31 0.0 A1000002 0.825269 0.398688 0.139396 0.843725 0.605194 0.406122 -0.128677 0.023810 0.00 0.423077
2 2018-09-30 0.0 A1000011 0.315406 0.629745 0.535854 0.197392 0.614416 0.320731 0.062660 0.023810 0.10 0.448718
3 2018-07-31 0.0 A10000481 0.002386 0.609360 0.366081 0.342243 0.870006 0.288692 0.078853 0.071429 0.05 0.179487
4 2018-07-31 0.0 A1000069 0.406310 0.405352 0.783015 0.563953 0.715454 0.512554 -0.261014 0.023810 0.00 0.423077
data.shape
(95806, 13)
# 看一下月份分布,我们用最后一个月做为跨时间验证集合
data.obs_mth.unique()
array(['2018-10-31', '2018-07-31', '2018-09-30', '2018-06-30',
       '2018-11-30'], dtype=object)
# 划分训练集和验证集
df_train = data[data.obs_mth != '2018-11-30'].reset_index().copy()
val = data[data.obs_mth == '2018-11-30'].reset_index().copy()
# 这是我们全部的变量,info结尾的是自己做的无监督系统输出的个人表现,score结尾的是收费的外部征信数据
lst = ['person_info','finance_info','credit_info','act_info','td_score','jxl_score','mj_score','rh_score']
df_train = df_train.sort_values(by = 'obs_mth',ascending = False)
df_train.head()
index obs_mth bad_ind uid td_score jxl_score mj_score rh_score zzc_score zcx_score person_info finance_info credit_info act_info
0 0 2018-10-31 0.0 A10000005 0.675349 0.144072 0.186899 0.483640 0.928328 0.369644 -0.322581 0.023810 0.00 0.217949
33407 33407 2018-10-31 0.0 A2810176 0.146055 0.079922 0.250568 0.045240 0.766906 0.413713 0.013863 0.023810 0.00 0.269231
33383 33383 2018-10-31 0.0 A2807687 0.551366 0.300781 0.225007 0.045447 0.735733 0.684182 -0.261014 0.071429 0.03 0.269231
33379 33379 2018-10-31 0.0 A2807232 0.708547 0.769513 0.928457 0.739716 0.947453 0.361551 -0.128677 0.047619 0.00 0.269231
33376 33376 2018-10-31 0.0 A2806932 0.482248 0.116658 0.286273 0.056618 0.047024 0.890433 0.078853 0.047619 0.00 0.269231
df_train = df_train.sort_values(by = 'obs_mth',ascending = False)

rank_lst = []
for i in range(1,len(df_train)+1):
    rank_lst.append(i)
    
df_train['rank'] = rank_lst

df_train['rank'] = df_train['rank']/len(df_train)

pct_lst = []
for x in df_train['rank']:
    if x <= 0.2:
        x = 1
    elif x <= 0.4:
        x = 2
    elif x <= 0.6:
        x = 3
    elif x <= 0.8:
        x = 4
    else:
        x = 5
    pct_lst.append(x)
df_train['rank'] = pct_lst        
# train = train.drop('obs_mth',axis = 1)
df_train.head()
index obs_mth bad_ind uid td_score jxl_score mj_score rh_score zzc_score zcx_score person_info finance_info credit_info act_info rank
0 0 2018-10-31 0.0 A10000005 0.675349 0.144072 0.186899 0.483640 0.928328 0.369644 -0.322581 0.023810 0.00 0.217949 1
33272 33272 2018-10-31 0.0 A2798022 0.448084 0.748679 0.328625 0.132682 0.167770 0.727771 -0.322581 0.023810 0.00 0.269231 1
33253 33253 2018-10-31 0.0 A2796636 0.129880 0.158756 0.240867 0.828666 0.232677 0.049676 -0.261014 0.071429 0.44 0.282051 1
32983 32983 2018-10-31 0.0 A2773761 0.644977 0.562769 0.751521 0.482639 0.486205 0.844002 0.078853 0.095238 0.03 0.282051 1
33369 33369 2018-10-31 0.0 A2806416 0.698203 0.164079 0.634027 0.013182 0.883847 0.174621 -0.322581 0.071429 0.06 0.269231 1
df_train['rank'].groupby(df_train['rank']).count()
rank
1    15966
2    15966
3    15966
4    15966
5    15967
Name: rank, dtype: int64

bin_record = pd.DataFrame()
for col in col_lst:
cb.pct_bin(data,col,'label')
cb.plot_woe()
data[col] = cb.trans_to_woe(data[col])
rcd = cb.get_bin_stats()
if bin_record.empty:
bin_record = rcd
else:
bin_record = bin_record.append(rcd)
bin_record.head(8)

len(df_train)
79831
# 定义lgb函数
def LGB_test(train_x,train_y,test_x,test_y):
    from multiprocessing import cpu_count
    clf = lgb.LGBMClassifier(
        boosting_type='gbdt', 
        num_leaves=31, 
        reg_alpha=0.0, 
        reg_lambda=1,
        max_depth=2, 
        n_estimators=800,
        max_features = 140, 
        objective='binary',
        subsample=0.7, 
        colsample_bytree=0.7, 
        subsample_freq=1,
        learning_rate=0.05, 
        min_child_weight=50,
        random_state=None,
        n_jobs=cpu_count()-1,
        num_iterations = 800 #迭代次数
    )
    clf.fit(train_x, train_y,eval_set=[(train_x, train_y),(test_x,test_y)],eval_metric='auc',early_stopping_rounds=100)
    print(clf.n_features_)

    return clf,clf.best_score_[ 'valid_1']['auc']
feature_lst = {}
ks_train_lst = []
ks_test_lst = []
for rk in set(df_train['rank']):   
    
    # 测试集8.18以后作为跨时间验证集
    
    # 定义模型训练集与测试集
    ttest = df_train[df_train['rank'] ==  rk]
    ttrain = df_train[df_train['rank'] !=  rk]
    
    train = ttrain[lst]
    train_y = ttrain.bad_ind
    
    test = ttest[lst]
    test_y = ttest.bad_ind    
    
    start = time.time()
    model,auc = LGB_test(train,train_y,test,test_y)                    
    end = time.time()
    
    # 模型贡献度放在feture中
    feature = pd.DataFrame(
                {'name' : model.booster_.feature_name(),
                'importance' : model.feature_importances_
              }).sort_values(by =  ['importance'],ascending = False)
    
       
    # 计算训练集、测试集、验证集上的KS和AUC

    y_pred_train_lgb = model.predict_proba(train)[:, 1]
    y_pred_test_lgb = model.predict_proba(test)[:, 1]


    train_fpr_lgb, train_tpr_lgb, _ = roc_curve(train_y, y_pred_train_lgb)
    test_fpr_lgb, test_tpr_lgb, _ = roc_curve(test_y, y_pred_test_lgb)


    train_ks = abs(train_fpr_lgb - train_tpr_lgb).max()
    test_ks = abs(test_fpr_lgb - test_tpr_lgb).max()


    train_auc = metrics.auc(train_fpr_lgb, train_tpr_lgb)
    test_auc = metrics.auc(test_fpr_lgb, test_tpr_lgb)
    
    ks_train_lst.append(train_ks)
    ks_test_lst.append(test_ks)    

    feature_lst[str(rk)] = feature[feature.importance>=20].name

train_ks = np.mean(ks_train_lst)
test_ks = np.mean(ks_test_lst)

ft_lst = {}
for i in range(1,6):
    ft_lst[str(i)] = feature_lst[str(i)]

fn_lst=list(set(ft_lst['1']) & set(ft_lst['2']) 
    & set(ft_lst['3']) & set(ft_lst['4']) &set(ft_lst['5']))

print('train_ks: ',train_ks)
print('test_ks: ',test_ks)

print('ft_lst: ',fn_lst )
E:\Anaconda3\envs\sklearn\lib\site-packages\lightgbm\engine.py:116: UserWarning: Found `num_iterations` in params. Will use it instead of argument
  warnings.warn("Found `{}` in params. Will use it instead of argument".format(alias))


[1]	training's auc: 0.671177	training's binary_logloss: 0.0837149	valid_1's auc: 0.593289	valid_1's binary_logloss: 0.122621
Training until validation scores don't improve for 100 rounds.
[2]	training's auc: 0.790094	training's binary_logloss: 0.0824434	valid_1's auc: 0.748081	valid_1's binary_logloss: 0.120209
[3]	training's auc: 0.794322	training's binary_logloss: 0.0819485	valid_1's auc: 0.769838	valid_1's binary_logloss: 0.119323
[4]	training's auc: 0.793401	training's binary_logloss: 0.08136	valid_1's auc: 0.766694	valid_1's binary_logloss: 0.11835
[5]	training's auc: 0.79442	training's binary_logloss: 0.0805306	valid_1's auc: 0.770955	valid_1's binary_logloss: 0.116824
[6]	training's auc: 0.799359	training's binary_logloss: 0.0797782	valid_1's auc: 0.777701	valid_1's binary_logloss: 0.115472
[7]	training's auc: 0.7986	training's binary_logloss: 0.0793654	valid_1's auc: 0.77792	valid_1's binary_logloss: 0.114819
[8]	training's auc: 0.80116	training's binary_logloss: 0.0787197	valid_1's auc: 0.779639	valid_1's binary_logloss: 0.113766
[9]	training's auc: 0.804811	training's binary_logloss: 0.0784944	valid_1's auc: 0.776231	valid_1's binary_logloss: 0.113544
[10]	training's auc: 0.803304	training's binary_logloss: 0.0781703	valid_1's auc: 0.774568	valid_1's binary_logloss: 0.113122
[11]	training's auc: 0.804755	training's binary_logloss: 0.0777157	valid_1's auc: 0.778525	valid_1's binary_logloss: 0.11235
[12]	training's auc: 0.80572	training's binary_logloss: 0.077203	valid_1's auc: 0.780386	valid_1's binary_logloss: 0.111528
[13]	training's auc: 0.805872	training's binary_logloss: 0.0769671	valid_1's auc: 0.780168	valid_1's binary_logloss: 0.111249
[14]	training's auc: 0.804747	training's binary_logloss: 0.0767342	valid_1's auc: 0.776101	valid_1's binary_logloss: 0.110985
[15]	training's auc: 0.806523	training's binary_logloss: 0.0763647	valid_1's auc: 0.776495	valid_1's binary_logloss: 0.110407
[16]	training's auc: 0.807263	training's binary_logloss: 0.0760476	valid_1's auc: 0.777377	valid_1's binary_logloss: 0.109954
[17]	training's auc: 0.807666	training's binary_logloss: 0.0758342	valid_1's auc: 0.778523	valid_1's binary_logloss: 0.109683
[18]	training's auc: 0.808026	training's binary_logloss: 0.0755496	valid_1's auc: 0.782706	valid_1's binary_logloss: 0.109231
[19]	training's auc: 0.807956	training's binary_logloss: 0.0752868	valid_1's auc: 0.782144	valid_1's binary_logloss: 0.108865
[20]	training's auc: 0.808187	training's binary_logloss: 0.07511	valid_1's auc: 0.781477	valid_1's binary_logloss: 0.10868
[21]	training's auc: 0.808443	training's binary_logloss: 0.0749012	valid_1's auc: 0.781449	valid_1's binary_logloss: 0.10841
[22]	training's auc: 0.80858	training's binary_logloss: 0.0746929	valid_1's auc: 0.781784	valid_1's binary_logloss: 0.108141
[23]	training's auc: 0.808537	training's binary_logloss: 0.0745555	valid_1's auc: 0.781804	valid_1's binary_logloss: 0.107996
[24]	training's auc: 0.808534	training's binary_logloss: 0.0743843	valid_1's auc: 0.782433	valid_1's binary_logloss: 0.107742
[25]	training's auc: 0.80871	training's binary_logloss: 0.074262	valid_1's auc: 0.782362	valid_1's binary_logloss: 0.10764
[26]	training's auc: 0.808796	training's binary_logloss: 0.0741263	valid_1's auc: 0.782014	valid_1's binary_logloss: 0.107489
[27]	training's auc: 0.809004	training's binary_logloss: 0.0739832	valid_1's auc: 0.783511	valid_1's binary_logloss: 0.107278
[28]	training's auc: 0.810629	training's binary_logloss: 0.0738869	valid_1's auc: 0.783018	valid_1's binary_logloss: 0.107224
[29]	training's auc: 0.811459	training's binary_logloss: 0.0737598	valid_1's auc: 0.783637	valid_1's binary_logloss: 0.107058
[30]	training's auc: 0.812163	training's binary_logloss: 0.0736609	valid_1's auc: 0.784984	valid_1's binary_logloss: 0.10695
[31]	training's auc: 0.811803	training's binary_logloss: 0.0735418	valid_1's auc: 0.784155	valid_1's binary_logloss: 0.106847
[32]	training's auc: 0.812089	training's binary_logloss: 0.0734718	valid_1's auc: 0.784006	valid_1's binary_logloss: 0.106825
[33]	training's auc: 0.81212	training's binary_logloss: 0.0733773	valid_1's auc: 0.784219	valid_1's binary_logloss: 0.106737
[34]	training's auc: 0.812082	training's binary_logloss: 0.0733098	valid_1's auc: 0.784139	valid_1's binary_logloss: 0.106721
[35]	training's auc: 0.812312	training's binary_logloss: 0.0732273	valid_1's auc: 0.783888	valid_1's binary_logloss: 0.106653
[36]	training's auc: 0.81221	training's binary_logloss: 0.0731534	valid_1's auc: 0.783731	valid_1's binary_logloss: 0.106591
[37]	training's auc: 0.81269	training's binary_logloss: 0.0730713	valid_1's auc: 0.784174	valid_1's binary_logloss: 0.106514
[38]	training's auc: 0.812679	training's binary_logloss: 0.072993	valid_1's auc: 0.78418	valid_1's binary_logloss: 0.106475
[39]	training's auc: 0.813185	training's binary_logloss: 0.0729372	valid_1's auc: 0.784127	valid_1's binary_logloss: 0.106471
[40]	training's auc: 0.813126	training's binary_logloss: 0.0728752	valid_1's auc: 0.784373	valid_1's binary_logloss: 0.106428
[41]	training's auc: 0.813399	training's binary_logloss: 0.0728172	valid_1's auc: 0.784511	valid_1's binary_logloss: 0.106386
[42]	training's auc: 0.813647	training's binary_logloss: 0.0727682	valid_1's auc: 0.784372	valid_1's binary_logloss: 0.106363
[43]	training's auc: 0.813684	training's binary_logloss: 0.0727066	valid_1's auc: 0.784718	valid_1's binary_logloss: 0.106314
[44]	training's auc: 0.813822	training's binary_logloss: 0.0726594	valid_1's auc: 0.7847	valid_1's binary_logloss: 0.106274
[45]	training's auc: 0.814282	training's binary_logloss: 0.0726142	valid_1's auc: 0.784598	valid_1's binary_logloss: 0.106242
[46]	training's auc: 0.814422	training's binary_logloss: 0.0725718	valid_1's auc: 0.784624	valid_1's binary_logloss: 0.106222
[47]	training's auc: 0.814488	training's binary_logloss: 0.0725269	valid_1's auc: 0.784789	valid_1's binary_logloss: 0.106203
[48]	training's auc: 0.814764	training's binary_logloss: 0.0724902	valid_1's auc: 0.785033	valid_1's binary_logloss: 0.106166
[49]	training's auc: 0.814857	training's binary_logloss: 0.0724551	valid_1's auc: 0.785419	valid_1's binary_logloss: 0.106146
[50]	training's auc: 0.814942	training's binary_logloss: 0.0724262	valid_1's auc: 0.785531	valid_1's binary_logloss: 0.106129
[51]	training's auc: 0.81514	training's binary_logloss: 0.0723931	valid_1's auc: 0.785477	valid_1's binary_logloss: 0.106135
[52]	training's auc: 0.815228	training's binary_logloss: 0.0723504	valid_1's auc: 0.785603	valid_1's binary_logloss: 0.106082
[53]	training's auc: 0.815361	training's binary_logloss: 0.0723331	valid_1's auc: 0.785824	valid_1's binary_logloss: 0.106057
[54]	training's auc: 0.815387	training's binary_logloss: 0.0723119	valid_1's auc: 0.78613	valid_1's binary_logloss: 0.105999
[55]	training's auc: 0.815514	training's binary_logloss: 0.0722825	valid_1's auc: 0.785812	valid_1's binary_logloss: 0.105995
[56]	training's auc: 0.81556	training's binary_logloss: 0.0722503	valid_1's auc: 0.785934	valid_1's binary_logloss: 0.105974
[57]	training's auc: 0.815588	training's binary_logloss: 0.0722242	valid_1's auc: 0.785955	valid_1's binary_logloss: 0.105976
[58]	training's auc: 0.815732	training's binary_logloss: 0.0721906	valid_1's auc: 0.78627	valid_1's binary_logloss: 0.105962
[59]	training's auc: 0.815706	training's binary_logloss: 0.0721687	valid_1's auc: 0.786258	valid_1's binary_logloss: 0.105951
[60]	training's auc: 0.815718	training's binary_logloss: 0.0721445	valid_1's auc: 0.786547	valid_1's binary_logloss: 0.105923
[61]	training's auc: 0.815734	training's binary_logloss: 0.0721223	valid_1's auc: 0.786568	valid_1's binary_logloss: 0.105923
[62]	training's auc: 0.815961	training's binary_logloss: 0.0720899	valid_1's auc: 0.786347	valid_1's binary_logloss: 0.105958
[63]	training's auc: 0.815941	training's binary_logloss: 0.072071	valid_1's auc: 0.786343	valid_1's binary_logloss: 0.105958
[64]	training's auc: 0.816068	training's binary_logloss: 0.0720548	valid_1's auc: 0.786568	valid_1's binary_logloss: 0.105948
[65]	training's auc: 0.81616	training's binary_logloss: 0.0720343	valid_1's auc: 0.786472	valid_1's binary_logloss: 0.105956
[66]	training's auc: 0.816269	training's binary_logloss: 0.0720188	valid_1's auc: 0.786689	valid_1's binary_logloss: 0.105938
[67]	training's auc: 0.816373	training's binary_logloss: 0.0719948	valid_1's auc: 0.786295	valid_1's binary_logloss: 0.105967
[68]	training's auc: 0.816425	training's binary_logloss: 0.0719716	valid_1's auc: 0.786729	valid_1's binary_logloss: 0.105935
[69]	training's auc: 0.816552	training's binary_logloss: 0.0719538	valid_1's auc: 0.786615	valid_1's binary_logloss: 0.105947
[70]	training's auc: 0.816446	training's binary_logloss: 0.071935	valid_1's auc: 0.786604	valid_1's binary_logloss: 0.105956
[71]	training's auc: 0.816411	training's binary_logloss: 0.0719215	valid_1's auc: 0.78663	valid_1's binary_logloss: 0.105947
[72]	training's auc: 0.816383	training's binary_logloss: 0.0719075	valid_1's auc: 0.786761	valid_1's binary_logloss: 0.105954
[73]	training's auc: 0.816402	training's binary_logloss: 0.071894	valid_1's auc: 0.786845	valid_1's binary_logloss: 0.105942
[74]	training's auc: 0.816524	training's binary_logloss: 0.0718747	valid_1's auc: 0.78676	valid_1's binary_logloss: 0.105952
[75]	training's auc: 0.816573	training's binary_logloss: 0.0718592	valid_1's auc: 0.786748	valid_1's binary_logloss: 0.105953
[76]	training's auc: 0.816671	training's binary_logloss: 0.0718444	valid_1's auc: 0.786642	valid_1's binary_logloss: 0.105978
[77]	training's auc: 0.816744	training's binary_logloss: 0.0718337	valid_1's auc: 0.786565	valid_1's binary_logloss: 0.105976
[78]	training's auc: 0.816932	training's binary_logloss: 0.0718212	valid_1's auc: 0.786543	valid_1's binary_logloss: 0.105991
[79]	training's auc: 0.816983	training's binary_logloss: 0.0718046	valid_1's auc: 0.78653	valid_1's binary_logloss: 0.105997
[80]	training's auc: 0.81722	training's binary_logloss: 0.0717917	valid_1's auc: 0.786595	valid_1's binary_logloss: 0.106016
[81]	training's auc: 0.817299	training's binary_logloss: 0.0717716	valid_1's auc: 0.786638	valid_1's binary_logloss: 0.106026
[82]	training's auc: 0.817325	training's binary_logloss: 0.0717643	valid_1's auc: 0.786818	valid_1's binary_logloss: 0.106014
[83]	training's auc: 0.817473	training's binary_logloss: 0.0717534	valid_1's auc: 0.786651	valid_1's binary_logloss: 0.106037
[84]	training's auc: 0.817696	training's binary_logloss: 0.0717374	valid_1's auc: 0.786501	valid_1's binary_logloss: 0.106077
[85]	training's auc: 0.817887	training's binary_logloss: 0.0717207	valid_1's auc: 0.786448	valid_1's binary_logloss: 0.106083
[86]	training's auc: 0.817972	training's binary_logloss: 0.0717088	valid_1's auc: 0.78615	valid_1's binary_logloss: 0.106111
[87]	training's auc: 0.81801	training's binary_logloss: 0.0716979	valid_1's auc: 0.786182	valid_1's binary_logloss: 0.106116
[88]	training's auc: 0.818042	training's binary_logloss: 0.0716852	valid_1's auc: 0.786034	valid_1's binary_logloss: 0.106152
[89]	training's auc: 0.818076	training's binary_logloss: 0.0716731	valid_1's auc: 0.785794	valid_1's binary_logloss: 0.10618
[90]	training's auc: 0.818089	training's binary_logloss: 0.0716662	valid_1's auc: 0.785778	valid_1's binary_logloss: 0.106185
[91]	training's auc: 0.818269	training's binary_logloss: 0.0716563	valid_1's auc: 0.785701	valid_1's binary_logloss: 0.106191
[92]	training's auc: 0.818316	training's binary_logloss: 0.0716505	valid_1's auc: 0.785694	valid_1's binary_logloss: 0.10619
[93]	training's auc: 0.818289	training's binary_logloss: 0.0716419	valid_1's auc: 0.7855	valid_1's binary_logloss: 0.106214
[94]	training's auc: 0.818372	training's binary_logloss: 0.0716302	valid_1's auc: 0.785272	valid_1's binary_logloss: 0.106236
[95]	training's auc: 0.818424	training's binary_logloss: 0.0716234	valid_1's auc: 0.785335	valid_1's binary_logloss: 0.106238
[96]	training's auc: 0.818421	training's binary_logloss: 0.0716175	valid_1's auc: 0.785293	valid_1's binary_logloss: 0.106244
[97]	training's auc: 0.818421	training's binary_logloss: 0.071611	valid_1's auc: 0.785156	valid_1's binary_logloss: 0.10626
[98]	training's auc: 0.818456	training's binary_logloss: 0.0716046	valid_1's auc: 0.78505	valid_1's binary_logloss: 0.10627
[99]	training's auc: 0.818469	training's binary_logloss: 0.0715997	valid_1's auc: 0.785019	valid_1's binary_logloss: 0.106281
[100]	training's auc: 0.818453	training's binary_logloss: 0.0715917	valid_1's auc: 0.785081	valid_1's binary_logloss: 0.106274
[101]	training's auc: 0.818465	training's binary_logloss: 0.0715882	valid_1's auc: 0.785049	valid_1's binary_logloss: 0.10628
[102]	training's auc: 0.818512	training's binary_logloss: 0.0715827	valid_1's auc: 0.785139	valid_1's binary_logloss: 0.106266
[103]	training's auc: 0.818635	training's binary_logloss: 0.0715739	valid_1's auc: 0.785213	valid_1's binary_logloss: 0.106275
[104]	training's auc: 0.818576	training's binary_logloss: 0.0715629	valid_1's auc: 0.785026	valid_1's binary_logloss: 0.106282
[105]	training's auc: 0.818638	training's binary_logloss: 0.0715508	valid_1's auc: 0.785035	valid_1's binary_logloss: 0.106293
[106]	training's auc: 0.818756	training's binary_logloss: 0.0715429	valid_1's auc: 0.785193	valid_1's binary_logloss: 0.10628
[107]	training's auc: 0.818777	training's binary_logloss: 0.0715386	valid_1's auc: 0.78513	valid_1's binary_logloss: 0.106284
[108]	training's auc: 0.81883	training's binary_logloss: 0.0715356	valid_1's auc: 0.784956	valid_1's binary_logloss: 0.106299
[109]	training's auc: 0.81888	training's binary_logloss: 0.0715297	valid_1's auc: 0.785038	valid_1's binary_logloss: 0.106295
[110]	training's auc: 0.818889	training's binary_logloss: 0.0715239	valid_1's auc: 0.78495	valid_1's binary_logloss: 0.106304
[111]	training's auc: 0.818948	training's binary_logloss: 0.0715186	valid_1's auc: 0.784973	valid_1's binary_logloss: 0.106305
[112]	training's auc: 0.819003	training's binary_logloss: 0.071514	valid_1's auc: 0.784907	valid_1's binary_logloss: 0.106323
[113]	training's auc: 0.819074	training's binary_logloss: 0.0715064	valid_1's auc: 0.784838	valid_1's binary_logloss: 0.106329
[114]	training's auc: 0.819102	training's binary_logloss: 0.0715031	valid_1's auc: 0.784872	valid_1's binary_logloss: 0.106339
[115]	training's auc: 0.819181	training's binary_logloss: 0.0714946	valid_1's auc: 0.784782	valid_1's binary_logloss: 0.106345
[116]	training's auc: 0.819211	training's binary_logloss: 0.0714871	valid_1's auc: 0.784749	valid_1's binary_logloss: 0.106341
[117]	training's auc: 0.819215	training's binary_logloss: 0.0714807	valid_1's auc: 0.784544	valid_1's binary_logloss: 0.106372
[118]	training's auc: 0.81929	training's binary_logloss: 0.0714729	valid_1's auc: 0.784592	valid_1's binary_logloss: 0.106371
[119]	training's auc: 0.819368	training's binary_logloss: 0.0714673	valid_1's auc: 0.784579	valid_1's binary_logloss: 0.10637
[120]	training's auc: 0.819413	training's binary_logloss: 0.0714624	valid_1's auc: 0.784451	valid_1's binary_logloss: 0.106387
[121]	training's auc: 0.819462	training's binary_logloss: 0.071455	valid_1's auc: 0.784657	valid_1's binary_logloss: 0.106364
[122]	training's auc: 0.819496	training's binary_logloss: 0.0714507	valid_1's auc: 0.784709	valid_1's binary_logloss: 0.106369
[123]	training's auc: 0.81958	training's binary_logloss: 0.0714443	valid_1's auc: 0.784624	valid_1's binary_logloss: 0.106382
[124]	training's auc: 0.819635	training's binary_logloss: 0.0714367	valid_1's auc: 0.784379	valid_1's binary_logloss: 0.106401
[125]	training's auc: 0.81967	training's binary_logloss: 0.0714324	valid_1's auc: 0.784237	valid_1's binary_logloss: 0.106417
[126]	training's auc: 0.819719	training's binary_logloss: 0.0714271	valid_1's auc: 0.784327	valid_1's binary_logloss: 0.106417
[127]	training's auc: 0.819772	training's binary_logloss: 0.0714205	valid_1's auc: 0.784247	valid_1's binary_logloss: 0.106427
[128]	training's auc: 0.819792	training's binary_logloss: 0.071417	valid_1's auc: 0.784253	valid_1's binary_logloss: 0.10643
[129]	training's auc: 0.819848	training's binary_logloss: 0.0714141	valid_1's auc: 0.783987	valid_1's binary_logloss: 0.10645
[130]	training's auc: 0.81994	training's binary_logloss: 0.0714057	valid_1's auc: 0.784056	valid_1's binary_logloss: 0.106441
[131]	training's auc: 0.81996	training's binary_logloss: 0.0714018	valid_1's auc: 0.784088	valid_1's binary_logloss: 0.106439
[132]	training's auc: 0.819989	training's binary_logloss: 0.0713864	valid_1's auc: 0.784202	valid_1's binary_logloss: 0.106422
[133]	training's auc: 0.820084	training's binary_logloss: 0.0713787	valid_1's auc: 0.784184	valid_1's binary_logloss: 0.106438
[134]	training's auc: 0.82018	training's binary_logloss: 0.0713551	valid_1's auc: 0.784649	valid_1's binary_logloss: 0.106403
[135]	training's auc: 0.820187	training's binary_logloss: 0.0713526	valid_1's auc: 0.784664	valid_1's binary_logloss: 0.106395
[136]	training's auc: 0.820236	training's binary_logloss: 0.0713442	valid_1's auc: 0.7846	valid_1's binary_logloss: 0.106402
[137]	training's auc: 0.820281	training's binary_logloss: 0.07134	valid_1's auc: 0.784467	valid_1's binary_logloss: 0.106425
[138]	training's auc: 0.82031	training's binary_logloss: 0.0713346	valid_1's auc: 0.784498	valid_1's binary_logloss: 0.106406
[139]	training's auc: 0.820342	training's binary_logloss: 0.071329	valid_1's auc: 0.784551	valid_1's binary_logloss: 0.106404
[140]	training's auc: 0.820391	training's binary_logloss: 0.0713228	valid_1's auc: 0.784492	valid_1's binary_logloss: 0.106409
[141]	training's auc: 0.820472	training's binary_logloss: 0.0713172	valid_1's auc: 0.784594	valid_1's binary_logloss: 0.106407
[142]	training's auc: 0.820561	training's binary_logloss: 0.0713113	valid_1's auc: 0.784403	valid_1's binary_logloss: 0.106431
[143]	training's auc: 0.820572	training's binary_logloss: 0.0713074	valid_1's auc: 0.784214	valid_1's binary_logloss: 0.106449
[144]	training's auc: 0.820633	training's binary_logloss: 0.0713037	valid_1's auc: 0.784233	valid_1's binary_logloss: 0.106444
[145]	training's auc: 0.820673	training's binary_logloss: 0.0712998	valid_1's auc: 0.78431	valid_1's binary_logloss: 0.106438
[146]	training's auc: 0.820693	training's binary_logloss: 0.071294	valid_1's auc: 0.784178	valid_1's binary_logloss: 0.106461
[147]	training's auc: 0.820742	training's binary_logloss: 0.0712881	valid_1's auc: 0.784215	valid_1's binary_logloss: 0.106464
[148]	training's auc: 0.820745	training's binary_logloss: 0.0712848	valid_1's auc: 0.784344	valid_1's binary_logloss: 0.106447
[149]	training's auc: 0.82079	training's binary_logloss: 0.0712808	valid_1's auc: 0.784314	valid_1's binary_logloss: 0.106459
[150]	training's auc: 0.820794	training's binary_logloss: 0.0712769	valid_1's auc: 0.784297	valid_1's binary_logloss: 0.106463
[151]	training's auc: 0.820821	training's binary_logloss: 0.0712722	valid_1's auc: 0.784535	valid_1's binary_logloss: 0.10644
[152]	training's auc: 0.820869	training's binary_logloss: 0.0712653	valid_1's auc: 0.78438	valid_1's binary_logloss: 0.106455
[153]	training's auc: 0.82085	training's binary_logloss: 0.0712634	valid_1's auc: 0.784364	valid_1's binary_logloss: 0.106461
[154]	training's auc: 0.82092	training's binary_logloss: 0.0712579	valid_1's auc: 0.784313	valid_1's binary_logloss: 0.106469
[155]	training's auc: 0.820927	training's binary_logloss: 0.0712541	valid_1's auc: 0.78425	valid_1's binary_logloss: 0.106471
[156]	training's auc: 0.821014	training's binary_logloss: 0.0712325	valid_1's auc: 0.784621	valid_1's binary_logloss: 0.106439
[157]	training's auc: 0.821049	training's binary_logloss: 0.0712301	valid_1's auc: 0.784657	valid_1's binary_logloss: 0.106439
[158]	training's auc: 0.821087	training's binary_logloss: 0.0712243	valid_1's auc: 0.784842	valid_1's binary_logloss: 0.106423
[159]	training's auc: 0.821106	training's binary_logloss: 0.0712181	valid_1's auc: 0.78503	valid_1's binary_logloss: 0.106405
[160]	training's auc: 0.821107	training's binary_logloss: 0.0712141	valid_1's auc: 0.784844	valid_1's binary_logloss: 0.106419
[161]	training's auc: 0.82119	training's binary_logloss: 0.0712073	valid_1's auc: 0.784637	valid_1's binary_logloss: 0.106434
Early stopping, best iteration is:
[61]	training's auc: 0.815734	training's binary_logloss: 0.0721223	valid_1's auc: 0.786568	valid_1's binary_logloss: 0.105923
8
[1]	training's auc: 0.665367	training's binary_logloss: 0.0868856	valid_1's auc: 0.638034	valid_1's binary_logloss: 0.109915
Training until validation scores don't improve for 100 rounds.
[2]	training's auc: 0.791507	training's binary_logloss: 0.0855856	valid_1's auc: 0.778478	valid_1's binary_logloss: 0.108111
[3]	training's auc: 0.795061	training's binary_logloss: 0.0850129	valid_1's auc: 0.782413	valid_1's binary_logloss: 0.107401
[4]	training's auc: 0.793479	training's binary_logloss: 0.0844315	valid_1's auc: 0.778869	valid_1's binary_logloss: 0.106679
[5]	training's auc: 0.801927	training's binary_logloss: 0.0834844	valid_1's auc: 0.787463	valid_1's binary_logloss: 0.10532
[6]	training's auc: 0.800946	training's binary_logloss: 0.0827504	valid_1's auc: 0.785723	valid_1's binary_logloss: 0.104338
[7]	training's auc: 0.800642	training's binary_logloss: 0.0822815	valid_1's auc: 0.785003	valid_1's binary_logloss: 0.103808
[8]	training's auc: 0.802571	training's binary_logloss: 0.0816004	valid_1's auc: 0.787851	valid_1's binary_logloss: 0.102893
[9]	training's auc: 0.804417	training's binary_logloss: 0.0814074	valid_1's auc: 0.789965	valid_1's binary_logloss: 0.102667
[10]	training's auc: 0.804715	training's binary_logloss: 0.0810815	valid_1's auc: 0.789315	valid_1's binary_logloss: 0.102308
[11]	training's auc: 0.80572	training's binary_logloss: 0.0805795	valid_1's auc: 0.790888	valid_1's binary_logloss: 0.101646
[12]	training's auc: 0.805909	training's binary_logloss: 0.0800685	valid_1's auc: 0.791102	valid_1's binary_logloss: 0.101022
[13]	training's auc: 0.805148	training's binary_logloss: 0.0798425	valid_1's auc: 0.789926	valid_1's binary_logloss: 0.100788
[14]	training's auc: 0.805025	training's binary_logloss: 0.0796195	valid_1's auc: 0.788701	valid_1's binary_logloss: 0.100546
[15]	training's auc: 0.806107	training's binary_logloss: 0.0792353	valid_1's auc: 0.790085	valid_1's binary_logloss: 0.100016
[16]	training's auc: 0.80629	training's binary_logloss: 0.0788918	valid_1's auc: 0.790386	valid_1's binary_logloss: 0.0995958
[17]	training's auc: 0.806259	training's binary_logloss: 0.0786638	valid_1's auc: 0.790254	valid_1's binary_logloss: 0.0993615
[18]	training's auc: 0.806691	training's binary_logloss: 0.0783631	valid_1's auc: 0.790354	valid_1's binary_logloss: 0.0990228
[19]	training's auc: 0.80746	training's binary_logloss: 0.0781036	valid_1's auc: 0.791374	valid_1's binary_logloss: 0.0987136
[20]	training's auc: 0.80775	training's binary_logloss: 0.0779183	valid_1's auc: 0.791057	valid_1's binary_logloss: 0.0985308
[21]	training's auc: 0.808059	training's binary_logloss: 0.0776847	valid_1's auc: 0.79161	valid_1's binary_logloss: 0.098227
[22]	training's auc: 0.807976	training's binary_logloss: 0.0774765	valid_1's auc: 0.791876	valid_1's binary_logloss: 0.0979735
[23]	training's auc: 0.808149	training's binary_logloss: 0.0773444	valid_1's auc: 0.792268	valid_1's binary_logloss: 0.0978359
[24]	training's auc: 0.808157	training's binary_logloss: 0.0771487	valid_1's auc: 0.791756	valid_1's binary_logloss: 0.0976358
[25]	training's auc: 0.808356	training's binary_logloss: 0.0770355	valid_1's auc: 0.791112	valid_1's binary_logloss: 0.0975479
[26]	training's auc: 0.808448	training's binary_logloss: 0.0768953	valid_1's auc: 0.791684	valid_1's binary_logloss: 0.097354
[27]	training's auc: 0.808386	training's binary_logloss: 0.0767449	valid_1's auc: 0.79175	valid_1's binary_logloss: 0.0972002
[28]	training's auc: 0.809099	training's binary_logloss: 0.0766654	valid_1's auc: 0.792508	valid_1's binary_logloss: 0.0971177
[29]	training's auc: 0.80913	training's binary_logloss: 0.0765278	valid_1's auc: 0.792793	valid_1's binary_logloss: 0.0969534
[30]	training's auc: 0.809349	training's binary_logloss: 0.0764214	valid_1's auc: 0.793102	valid_1's binary_logloss: 0.0968668
[31]	training's auc: 0.809617	training's binary_logloss: 0.0762925	valid_1's auc: 0.793769	valid_1's binary_logloss: 0.0967063
[32]	training's auc: 0.809557	training's binary_logloss: 0.076225	valid_1's auc: 0.793807	valid_1's binary_logloss: 0.0966626
[33]	training's auc: 0.810126	training's binary_logloss: 0.0761187	valid_1's auc: 0.794545	valid_1's binary_logloss: 0.0965436
[34]	training's auc: 0.809974	training's binary_logloss: 0.0760628	valid_1's auc: 0.794193	valid_1's binary_logloss: 0.0964899
[35]	training's auc: 0.810479	training's binary_logloss: 0.0759813	valid_1's auc: 0.794976	valid_1's binary_logloss: 0.0963861
[36]	training's auc: 0.810464	training's binary_logloss: 0.075896	valid_1's auc: 0.795457	valid_1's binary_logloss: 0.0962777
[37]	training's auc: 0.810701	training's binary_logloss: 0.0758081	valid_1's auc: 0.795639	valid_1's binary_logloss: 0.0961879
[38]	training's auc: 0.810869	training's binary_logloss: 0.0757245	valid_1's auc: 0.795936	valid_1's binary_logloss: 0.0961019
[39]	training's auc: 0.811406	training's binary_logloss: 0.0756708	valid_1's auc: 0.796704	valid_1's binary_logloss: 0.0960417
[40]	training's auc: 0.811464	training's binary_logloss: 0.0756043	valid_1's auc: 0.797168	valid_1's binary_logloss: 0.095968
[41]	training's auc: 0.811443	training's binary_logloss: 0.0755448	valid_1's auc: 0.797791	valid_1's binary_logloss: 0.0958963
[42]	training's auc: 0.81172	training's binary_logloss: 0.0754894	valid_1's auc: 0.798391	valid_1's binary_logloss: 0.09584
[43]	training's auc: 0.811679	training's binary_logloss: 0.0754119	valid_1's auc: 0.79814	valid_1's binary_logloss: 0.0957827
[44]	training's auc: 0.811701	training's binary_logloss: 0.0753675	valid_1's auc: 0.798242	valid_1's binary_logloss: 0.0957432
[45]	training's auc: 0.81173	training's binary_logloss: 0.0753183	valid_1's auc: 0.798303	valid_1's binary_logloss: 0.0957024
[46]	training's auc: 0.811994	training's binary_logloss: 0.0752779	valid_1's auc: 0.798184	valid_1's binary_logloss: 0.0956843
[47]	training's auc: 0.811946	training's binary_logloss: 0.0752103	valid_1's auc: 0.798203	valid_1's binary_logloss: 0.0956371
[48]	training's auc: 0.812328	training's binary_logloss: 0.0751747	valid_1's auc: 0.798574	valid_1's binary_logloss: 0.0955971
[49]	training's auc: 0.81253	training's binary_logloss: 0.0751401	valid_1's auc: 0.798755	valid_1's binary_logloss: 0.0955682
[50]	training's auc: 0.812567	training's binary_logloss: 0.075102	valid_1's auc: 0.798895	valid_1's binary_logloss: 0.0955279
[51]	training's auc: 0.812545	training's binary_logloss: 0.0750742	valid_1's auc: 0.798918	valid_1's binary_logloss: 0.0955016
[52]	training's auc: 0.812577	training's binary_logloss: 0.0750226	valid_1's auc: 0.798909	valid_1's binary_logloss: 0.095471
[53]	training's auc: 0.812859	training's binary_logloss: 0.0750067	valid_1's auc: 0.798986	valid_1's binary_logloss: 0.0954679
[54]	training's auc: 0.8129	training's binary_logloss: 0.0749863	valid_1's auc: 0.799125	valid_1's binary_logloss: 0.0954408
[55]	training's auc: 0.813031	training's binary_logloss: 0.0749588	valid_1's auc: 0.799384	valid_1's binary_logloss: 0.0954081
[56]	training's auc: 0.812881	training's binary_logloss: 0.074927	valid_1's auc: 0.799338	valid_1's binary_logloss: 0.0953942
[57]	training's auc: 0.812888	training's binary_logloss: 0.0748955	valid_1's auc: 0.799291	valid_1's binary_logloss: 0.0953836
[58]	training's auc: 0.813015	training's binary_logloss: 0.0748659	valid_1's auc: 0.799475	valid_1's binary_logloss: 0.0953671
[59]	training's auc: 0.813111	training's binary_logloss: 0.0748398	valid_1's auc: 0.799842	valid_1's binary_logloss: 0.0953486
[60]	training's auc: 0.813214	training's binary_logloss: 0.0748196	valid_1's auc: 0.800158	valid_1's binary_logloss: 0.0953262
[61]	training's auc: 0.813369	training's binary_logloss: 0.0747917	valid_1's auc: 0.80028	valid_1's binary_logloss: 0.0953142
[62]	training's auc: 0.813347	training's binary_logloss: 0.074751	valid_1's auc: 0.800148	valid_1's binary_logloss: 0.0952901
[63]	training's auc: 0.813447	training's binary_logloss: 0.0747295	valid_1's auc: 0.800345	valid_1's binary_logloss: 0.0952695
[64]	training's auc: 0.813481	training's binary_logloss: 0.0747062	valid_1's auc: 0.80039	valid_1's binary_logloss: 0.0952386
[65]	training's auc: 0.813693	training's binary_logloss: 0.0746838	valid_1's auc: 0.800635	valid_1's binary_logloss: 0.0952214
[66]	training's auc: 0.81388	training's binary_logloss: 0.0746691	valid_1's auc: 0.800462	valid_1's binary_logloss: 0.0952094
[67]	training's auc: 0.814062	training's binary_logloss: 0.0746476	valid_1's auc: 0.800544	valid_1's binary_logloss: 0.0951982
[68]	training's auc: 0.814117	training's binary_logloss: 0.074624	valid_1's auc: 0.800664	valid_1's binary_logloss: 0.0951826
[69]	training's auc: 0.814249	training's binary_logloss: 0.0746072	valid_1's auc: 0.800589	valid_1's binary_logloss: 0.0951832
[70]	training's auc: 0.814288	training's binary_logloss: 0.0745909	valid_1's auc: 0.800656	valid_1's binary_logloss: 0.0951881
[71]	training's auc: 0.814359	training's binary_logloss: 0.074577	valid_1's auc: 0.800828	valid_1's binary_logloss: 0.0951731
[72]	training's auc: 0.814241	training's binary_logloss: 0.0745619	valid_1's auc: 0.80094	valid_1's binary_logloss: 0.0951546
[73]	training's auc: 0.814325	training's binary_logloss: 0.0745471	valid_1's auc: 0.800995	valid_1's binary_logloss: 0.0951633
[74]	training's auc: 0.81452	training's binary_logloss: 0.0745357	valid_1's auc: 0.801002	valid_1's binary_logloss: 0.0951605
[75]	training's auc: 0.814656	training's binary_logloss: 0.0745232	valid_1's auc: 0.801098	valid_1's binary_logloss: 0.095154
[76]	training's auc: 0.814743	training's binary_logloss: 0.0745071	valid_1's auc: 0.801275	valid_1's binary_logloss: 0.0951344
[77]	training's auc: 0.81473	training's binary_logloss: 0.0744907	valid_1's auc: 0.80116	valid_1's binary_logloss: 0.0951438
[78]	training's auc: 0.814767	training's binary_logloss: 0.0744837	valid_1's auc: 0.801177	valid_1's binary_logloss: 0.0951393
[79]	training's auc: 0.814784	training's binary_logloss: 0.0744692	valid_1's auc: 0.801294	valid_1's binary_logloss: 0.0951311
[80]	training's auc: 0.814813	training's binary_logloss: 0.0744599	valid_1's auc: 0.801402	valid_1's binary_logloss: 0.0951247
[81]	training's auc: 0.814974	training's binary_logloss: 0.0744446	valid_1's auc: 0.801533	valid_1's binary_logloss: 0.0951186
[82]	training's auc: 0.815008	training's binary_logloss: 0.074433	valid_1's auc: 0.801598	valid_1's binary_logloss: 0.0951093
[83]	training's auc: 0.81512	training's binary_logloss: 0.0744136	valid_1's auc: 0.801794	valid_1's binary_logloss: 0.0950982
[84]	training's auc: 0.815284	training's binary_logloss: 0.0744	valid_1's auc: 0.801921	valid_1's binary_logloss: 0.0951031
[85]	training's auc: 0.81543	training's binary_logloss: 0.0743888	valid_1's auc: 0.801823	valid_1's binary_logloss: 0.0951
[86]	training's auc: 0.815413	training's binary_logloss: 0.0743766	valid_1's auc: 0.801803	valid_1's binary_logloss: 0.0950919
[87]	training's auc: 0.815411	training's binary_logloss: 0.0743599	valid_1's auc: 0.80168	valid_1's binary_logloss: 0.0950894
[88]	training's auc: 0.815474	training's binary_logloss: 0.0743497	valid_1's auc: 0.80175	valid_1's binary_logloss: 0.0950732
[89]	training's auc: 0.815529	training's binary_logloss: 0.0743353	valid_1's auc: 0.801893	valid_1's binary_logloss: 0.0950631
[90]	training's auc: 0.815556	training's binary_logloss: 0.0743267	valid_1's auc: 0.801935	valid_1's binary_logloss: 0.0950573
[91]	training's auc: 0.815756	training's binary_logloss: 0.0743137	valid_1's auc: 0.801872	valid_1's binary_logloss: 0.0950584
[92]	training's auc: 0.81582	training's binary_logloss: 0.0743069	valid_1's auc: 0.801957	valid_1's binary_logloss: 0.0950565
[93]	training's auc: 0.815822	training's binary_logloss: 0.0742966	valid_1's auc: 0.802232	valid_1's binary_logloss: 0.0950399
[94]	training's auc: 0.815967	training's binary_logloss: 0.0742838	valid_1's auc: 0.802182	valid_1's binary_logloss: 0.0950338
[95]	training's auc: 0.816057	training's binary_logloss: 0.0742736	valid_1's auc: 0.802218	valid_1's binary_logloss: 0.0950291
[96]	training's auc: 0.816071	training's binary_logloss: 0.0742607	valid_1's auc: 0.802331	valid_1's binary_logloss: 0.0950074
[97]	training's auc: 0.81615	training's binary_logloss: 0.074253	valid_1's auc: 0.802446	valid_1's binary_logloss: 0.095004
[98]	training's auc: 0.8161	training's binary_logloss: 0.074242	valid_1's auc: 0.802353	valid_1's binary_logloss: 0.0950067
[99]	training's auc: 0.816152	training's binary_logloss: 0.0742354	valid_1's auc: 0.802258	valid_1's binary_logloss: 0.0950091
[100]	training's auc: 0.816145	training's binary_logloss: 0.0742291	valid_1's auc: 0.802428	valid_1's binary_logloss: 0.0949979
[101]	training's auc: 0.816168	training's binary_logloss: 0.0742227	valid_1's auc: 0.80237	valid_1's binary_logloss: 0.0950027
[102]	training's auc: 0.81624	training's binary_logloss: 0.074216	valid_1's auc: 0.802414	valid_1's binary_logloss: 0.0949976
[103]	training's auc: 0.81639	training's binary_logloss: 0.074204	valid_1's auc: 0.802322	valid_1's binary_logloss: 0.095003
[104]	training's auc: 0.81646	training's binary_logloss: 0.074193	valid_1's auc: 0.802501	valid_1's binary_logloss: 0.0949953
[105]	training's auc: 0.816562	training's binary_logloss: 0.0741845	valid_1's auc: 0.802451	valid_1's binary_logloss: 0.0949929
[106]	training's auc: 0.816644	training's binary_logloss: 0.0741798	valid_1's auc: 0.802444	valid_1's binary_logloss: 0.0949924
[107]	training's auc: 0.816782	training's binary_logloss: 0.0741715	valid_1's auc: 0.802285	valid_1's binary_logloss: 0.0949931
[108]	training's auc: 0.816853	training's binary_logloss: 0.0741645	valid_1's auc: 0.802388	valid_1's binary_logloss: 0.0949899
[109]	training's auc: 0.816929	training's binary_logloss: 0.0741558	valid_1's auc: 0.802352	valid_1's binary_logloss: 0.09499
[110]	training's auc: 0.817009	training's binary_logloss: 0.0741498	valid_1's auc: 0.802288	valid_1's binary_logloss: 0.0949951
[111]	training's auc: 0.817059	training's binary_logloss: 0.0741419	valid_1's auc: 0.802315	valid_1's binary_logloss: 0.094996
[112]	training's auc: 0.817057	training's binary_logloss: 0.0741364	valid_1's auc: 0.802221	valid_1's binary_logloss: 0.0950033
[113]	training's auc: 0.817085	training's binary_logloss: 0.0741058	valid_1's auc: 0.802134	valid_1's binary_logloss: 0.0949912
[114]	training's auc: 0.817123	training's binary_logloss: 0.0741015	valid_1's auc: 0.802173	valid_1's binary_logloss: 0.0949877
[115]	training's auc: 0.817113	training's binary_logloss: 0.0740943	valid_1's auc: 0.802096	valid_1's binary_logloss: 0.0949946
[116]	training's auc: 0.817139	training's binary_logloss: 0.0740871	valid_1's auc: 0.80213	valid_1's binary_logloss: 0.094992
[117]	training's auc: 0.817159	training's binary_logloss: 0.0740785	valid_1's auc: 0.802129	valid_1's binary_logloss: 0.0949948
[118]	training's auc: 0.817272	training's binary_logloss: 0.0740693	valid_1's auc: 0.802091	valid_1's binary_logloss: 0.0949977
[119]	training's auc: 0.817387	training's binary_logloss: 0.0740614	valid_1's auc: 0.801905	valid_1's binary_logloss: 0.0950029
[120]	training's auc: 0.817514	training's binary_logloss: 0.0740524	valid_1's auc: 0.801835	valid_1's binary_logloss: 0.0950065
[121]	training's auc: 0.817639	training's binary_logloss: 0.074043	valid_1's auc: 0.801834	valid_1's binary_logloss: 0.0950134
[122]	training's auc: 0.817786	training's binary_logloss: 0.0740341	valid_1's auc: 0.801934	valid_1's binary_logloss: 0.0950115
[123]	training's auc: 0.817819	training's binary_logloss: 0.074026	valid_1's auc: 0.801927	valid_1's binary_logloss: 0.0950144
[124]	training's auc: 0.817866	training's binary_logloss: 0.0740188	valid_1's auc: 0.801899	valid_1's binary_logloss: 0.0950179
[125]	training's auc: 0.817929	training's binary_logloss: 0.0740085	valid_1's auc: 0.801989	valid_1's binary_logloss: 0.0950056
[126]	training's auc: 0.817945	training's binary_logloss: 0.0740053	valid_1's auc: 0.802016	valid_1's binary_logloss: 0.0950052
[127]	training's auc: 0.817971	training's binary_logloss: 0.0740011	valid_1's auc: 0.801981	valid_1's binary_logloss: 0.0950142
[128]	training's auc: 0.818047	training's binary_logloss: 0.0739978	valid_1's auc: 0.801979	valid_1's binary_logloss: 0.0950113
[129]	training's auc: 0.818142	training's binary_logloss: 0.073992	valid_1's auc: 0.80193	valid_1's binary_logloss: 0.095014
[130]	training's auc: 0.818194	training's binary_logloss: 0.073985	valid_1's auc: 0.801948	valid_1's binary_logloss: 0.0950175
[131]	training's auc: 0.818254	training's binary_logloss: 0.0739822	valid_1's auc: 0.801862	valid_1's binary_logloss: 0.0950211
[132]	training's auc: 0.818281	training's binary_logloss: 0.0739793	valid_1's auc: 0.801814	valid_1's binary_logloss: 0.0950347
[133]	training's auc: 0.818395	training's binary_logloss: 0.0739714	valid_1's auc: 0.80174	valid_1's binary_logloss: 0.0950386
[134]	training's auc: 0.818446	training's binary_logloss: 0.0739676	valid_1's auc: 0.801676	valid_1's binary_logloss: 0.0950468
[135]	training's auc: 0.818497	training's binary_logloss: 0.0739609	valid_1's auc: 0.801554	valid_1's binary_logloss: 0.0950441
[136]	training's auc: 0.818522	training's binary_logloss: 0.0739569	valid_1's auc: 0.801637	valid_1's binary_logloss: 0.0950404
[137]	training's auc: 0.81855	training's binary_logloss: 0.0739527	valid_1's auc: 0.801666	valid_1's binary_logloss: 0.09504
[138]	training's auc: 0.818545	training's binary_logloss: 0.073948	valid_1's auc: 0.801624	valid_1's binary_logloss: 0.095049
[139]	training's auc: 0.818598	training's binary_logloss: 0.0739444	valid_1's auc: 0.801586	valid_1's binary_logloss: 0.0950465
[140]	training's auc: 0.818629	training's binary_logloss: 0.0739346	valid_1's auc: 0.801645	valid_1's binary_logloss: 0.0950355
[141]	training's auc: 0.818692	training's binary_logloss: 0.0739312	valid_1's auc: 0.801564	valid_1's binary_logloss: 0.0950512
[142]	training's auc: 0.818662	training's binary_logloss: 0.0739265	valid_1's auc: 0.801566	valid_1's binary_logloss: 0.0950428
[143]	training's auc: 0.818677	training's binary_logloss: 0.0739241	valid_1's auc: 0.80156	valid_1's binary_logloss: 0.0950412
[144]	training's auc: 0.81866	training's binary_logloss: 0.0739217	valid_1's auc: 0.801581	valid_1's binary_logloss: 0.0950383
[145]	training's auc: 0.81869	training's binary_logloss: 0.073917	valid_1's auc: 0.801534	valid_1's binary_logloss: 0.0950444
[146]	training's auc: 0.818702	training's binary_logloss: 0.0739142	valid_1's auc: 0.80157	valid_1's binary_logloss: 0.0950473
[147]	training's auc: 0.818768	training's binary_logloss: 0.0739062	valid_1's auc: 0.801601	valid_1's binary_logloss: 0.0950422
[148]	training's auc: 0.818787	training's binary_logloss: 0.073903	valid_1's auc: 0.801642	valid_1's binary_logloss: 0.095036
[149]	training's auc: 0.818727	training's binary_logloss: 0.0738935	valid_1's auc: 0.801648	valid_1's binary_logloss: 0.0950198
[150]	training's auc: 0.818757	training's binary_logloss: 0.0738884	valid_1's auc: 0.80171	valid_1's binary_logloss: 0.0950195
[151]	training's auc: 0.81883	training's binary_logloss: 0.0738825	valid_1's auc: 0.801704	valid_1's binary_logloss: 0.0950253
[152]	training's auc: 0.818895	training's binary_logloss: 0.073878	valid_1's auc: 0.801701	valid_1's binary_logloss: 0.0950301
[153]	training's auc: 0.818909	training's binary_logloss: 0.0738754	valid_1's auc: 0.801637	valid_1's binary_logloss: 0.0950361
[154]	training's auc: 0.818953	training's binary_logloss: 0.0738718	valid_1's auc: 0.801606	valid_1's binary_logloss: 0.0950366
[155]	training's auc: 0.819025	training's binary_logloss: 0.0738668	valid_1's auc: 0.801601	valid_1's binary_logloss: 0.0950417
[156]	training's auc: 0.819068	training's binary_logloss: 0.0738627	valid_1's auc: 0.801585	valid_1's binary_logloss: 0.0950482
[157]	training's auc: 0.819171	training's binary_logloss: 0.0738559	valid_1's auc: 0.801686	valid_1's binary_logloss: 0.0950449
[158]	training's auc: 0.819191	training's binary_logloss: 0.073852	valid_1's auc: 0.801697	valid_1's binary_logloss: 0.09504
[159]	training's auc: 0.819194	training's binary_logloss: 0.073848	valid_1's auc: 0.801609	valid_1's binary_logloss: 0.0950393
[160]	training's auc: 0.819294	training's binary_logloss: 0.073842	valid_1's auc: 0.801457	valid_1's binary_logloss: 0.0950461
[161]	training's auc: 0.819357	training's binary_logloss: 0.0738356	valid_1's auc: 0.801522	valid_1's binary_logloss: 0.0950438
[162]	training's auc: 0.819381	training's binary_logloss: 0.0738313	valid_1's auc: 0.801544	valid_1's binary_logloss: 0.0950454
[163]	training's auc: 0.819433	training's binary_logloss: 0.0738278	valid_1's auc: 0.801594	valid_1's binary_logloss: 0.0950475
[164]	training's auc: 0.819479	training's binary_logloss: 0.0738238	valid_1's auc: 0.801506	valid_1's binary_logloss: 0.0950578
[165]	training's auc: 0.819485	training's binary_logloss: 0.0738228	valid_1's auc: 0.801513	valid_1's binary_logloss: 0.0950475
[166]	training's auc: 0.819515	training's binary_logloss: 0.0738183	valid_1's auc: 0.801514	valid_1's binary_logloss: 0.0950485
[167]	training's auc: 0.819536	training's binary_logloss: 0.0738119	valid_1's auc: 0.801463	valid_1's binary_logloss: 0.0950497
[168]	training's auc: 0.819537	training's binary_logloss: 0.0738086	valid_1's auc: 0.801461	valid_1's binary_logloss: 0.09505
[169]	training's auc: 0.819573	training's binary_logloss: 0.0738042	valid_1's auc: 0.801511	valid_1's binary_logloss: 0.0950516
[170]	training's auc: 0.819603	training's binary_logloss: 0.0738014	valid_1's auc: 0.801525	valid_1's binary_logloss: 0.0950523
[171]	training's auc: 0.819669	training's binary_logloss: 0.0737929	valid_1's auc: 0.80155	valid_1's binary_logloss: 0.0950544
[172]	training's auc: 0.819661	training's binary_logloss: 0.073789	valid_1's auc: 0.801575	valid_1's binary_logloss: 0.0950543
[173]	training's auc: 0.819682	training's binary_logloss: 0.0737863	valid_1's auc: 0.801548	valid_1's binary_logloss: 0.0950514
[174]	training's auc: 0.819686	training's binary_logloss: 0.0737827	valid_1's auc: 0.80149	valid_1's binary_logloss: 0.0950591
[175]	training's auc: 0.819784	training's binary_logloss: 0.0737754	valid_1's auc: 0.801542	valid_1's binary_logloss: 0.0950543
[176]	training's auc: 0.819856	training's binary_logloss: 0.07377	valid_1's auc: 0.801415	valid_1's binary_logloss: 0.0950595
[177]	training's auc: 0.819883	training's binary_logloss: 0.0737671	valid_1's auc: 0.801387	valid_1's binary_logloss: 0.0950631
[178]	training's auc: 0.819954	training's binary_logloss: 0.0737635	valid_1's auc: 0.801343	valid_1's binary_logloss: 0.0950697
[179]	training's auc: 0.819976	training's binary_logloss: 0.0737608	valid_1's auc: 0.801329	valid_1's binary_logloss: 0.0950714
[180]	training's auc: 0.820057	training's binary_logloss: 0.0737549	valid_1's auc: 0.801323	valid_1's binary_logloss: 0.0950687
[181]	training's auc: 0.820072	training's binary_logloss: 0.0737527	valid_1's auc: 0.801345	valid_1's binary_logloss: 0.0950752
[182]	training's auc: 0.820129	training's binary_logloss: 0.0737491	valid_1's auc: 0.801278	valid_1's binary_logloss: 0.0950773
[183]	training's auc: 0.820178	training's binary_logloss: 0.0737455	valid_1's auc: 0.80128	valid_1's binary_logloss: 0.0950661
[184]	training's auc: 0.820237	training's binary_logloss: 0.073741	valid_1's auc: 0.801223	valid_1's binary_logloss: 0.0950696
[185]	training's auc: 0.820263	training's binary_logloss: 0.0737363	valid_1's auc: 0.801297	valid_1's binary_logloss: 0.095063
[186]	training's auc: 0.820275	training's binary_logloss: 0.0737336	valid_1's auc: 0.80128	valid_1's binary_logloss: 0.0950628
[187]	training's auc: 0.820333	training's binary_logloss: 0.0737266	valid_1's auc: 0.801124	valid_1's binary_logloss: 0.0950909
[188]	training's auc: 0.820422	training's binary_logloss: 0.073719	valid_1's auc: 0.801112	valid_1's binary_logloss: 0.0950951
[189]	training's auc: 0.820417	training's binary_logloss: 0.0737164	valid_1's auc: 0.801178	valid_1's binary_logloss: 0.0950911
[190]	training's auc: 0.820463	training's binary_logloss: 0.0737107	valid_1's auc: 0.801299	valid_1's binary_logloss: 0.0950894
[191]	training's auc: 0.820474	training's binary_logloss: 0.0737078	valid_1's auc: 0.801324	valid_1's binary_logloss: 0.0950889
[192]	training's auc: 0.820482	training's binary_logloss: 0.0737056	valid_1's auc: 0.801229	valid_1's binary_logloss: 0.0950925
[193]	training's auc: 0.820573	training's binary_logloss: 0.0736989	valid_1's auc: 0.801189	valid_1's binary_logloss: 0.0951055
[194]	training's auc: 0.820602	training's binary_logloss: 0.0736934	valid_1's auc: 0.800876	valid_1's binary_logloss: 0.0951183
[195]	training's auc: 0.820616	training's binary_logloss: 0.0736868	valid_1's auc: 0.800814	valid_1's binary_logloss: 0.0951133
[196]	training's auc: 0.820633	training's binary_logloss: 0.0736837	valid_1's auc: 0.80082	valid_1's binary_logloss: 0.0951169
[197]	training's auc: 0.820661	training's binary_logloss: 0.0736819	valid_1's auc: 0.800804	valid_1's binary_logloss: 0.0951116
[198]	training's auc: 0.820649	training's binary_logloss: 0.0736789	valid_1's auc: 0.800729	valid_1's binary_logloss: 0.0951241
[199]	training's auc: 0.820663	training's binary_logloss: 0.0736741	valid_1's auc: 0.800617	valid_1's binary_logloss: 0.0951238
[200]	training's auc: 0.820727	training's binary_logloss: 0.073668	valid_1's auc: 0.800694	valid_1's binary_logloss: 0.0951257
[201]	training's auc: 0.820767	training's binary_logloss: 0.0736626	valid_1's auc: 0.800984	valid_1's binary_logloss: 0.0951079
[202]	training's auc: 0.820804	training's binary_logloss: 0.0736584	valid_1's auc: 0.801181	valid_1's binary_logloss: 0.0950977
[203]	training's auc: 0.820833	training's binary_logloss: 0.0736564	valid_1's auc: 0.801223	valid_1's binary_logloss: 0.0950869
[204]	training's auc: 0.820883	training's binary_logloss: 0.0736367	valid_1's auc: 0.801209	valid_1's binary_logloss: 0.0950887
Early stopping, best iteration is:
[104]	training's auc: 0.81646	training's binary_logloss: 0.074193	valid_1's auc: 0.802501	valid_1's binary_logloss: 0.0949953
8
[1]	training's auc: 0.659021	training's binary_logloss: 0.0977451	valid_1's auc: 0.667922	valid_1's binary_logloss: 0.0667753
Training until validation scores don't improve for 100 rounds.
[2]	training's auc: 0.782929	training's binary_logloss: 0.0962916	valid_1's auc: 0.784101	valid_1's binary_logloss: 0.0659222
[3]	training's auc: 0.78618	training's binary_logloss: 0.0956767	valid_1's auc: 0.788715	valid_1's binary_logloss: 0.0655366
[4]	training's auc: 0.785892	training's binary_logloss: 0.095076	valid_1's auc: 0.790152	valid_1's binary_logloss: 0.0651196
[5]	training's auc: 0.792838	training's binary_logloss: 0.0940868	valid_1's auc: 0.800957	valid_1's binary_logloss: 0.0644159
[6]	training's auc: 0.797256	training's binary_logloss: 0.0932581	valid_1's auc: 0.806857	valid_1's binary_logloss: 0.063861
[7]	training's auc: 0.796345	training's binary_logloss: 0.0927891	valid_1's auc: 0.80429	valid_1's binary_logloss: 0.0635305
[8]	training's auc: 0.797244	training's binary_logloss: 0.0920677	valid_1's auc: 0.804323	valid_1's binary_logloss: 0.0630112
[9]	training's auc: 0.799111	training's binary_logloss: 0.0918718	valid_1's auc: 0.807082	valid_1's binary_logloss: 0.0628553
[10]	training's auc: 0.798274	training's binary_logloss: 0.091531	valid_1's auc: 0.805748	valid_1's binary_logloss: 0.0625961
[11]	training's auc: 0.798767	training's binary_logloss: 0.0909135	valid_1's auc: 0.806574	valid_1's binary_logloss: 0.0621389
[12]	training's auc: 0.801005	training's binary_logloss: 0.0903573	valid_1's auc: 0.8077	valid_1's binary_logloss: 0.0616888
[13]	training's auc: 0.800594	training's binary_logloss: 0.0901254	valid_1's auc: 0.807376	valid_1's binary_logloss: 0.0615163
[14]	training's auc: 0.799391	training's binary_logloss: 0.0898873	valid_1's auc: 0.804602	valid_1's binary_logloss: 0.0613258
[15]	training's auc: 0.799973	training's binary_logloss: 0.0894598	valid_1's auc: 0.80734	valid_1's binary_logloss: 0.0609837
[16]	training's auc: 0.800826	training's binary_logloss: 0.0890624	valid_1's auc: 0.807424	valid_1's binary_logloss: 0.0606438
[17]	training's auc: 0.800556	training's binary_logloss: 0.0888349	valid_1's auc: 0.807063	valid_1's binary_logloss: 0.0604495
[18]	training's auc: 0.800736	training's binary_logloss: 0.0885066	valid_1's auc: 0.808055	valid_1's binary_logloss: 0.0601512
[19]	training's auc: 0.801701	training's binary_logloss: 0.0882099	valid_1's auc: 0.809107	valid_1's binary_logloss: 0.0598891
[20]	training's auc: 0.801487	training's binary_logloss: 0.0880168	valid_1's auc: 0.80943	valid_1's binary_logloss: 0.0597102
[21]	training's auc: 0.801994	training's binary_logloss: 0.0877765	valid_1's auc: 0.810201	valid_1's binary_logloss: 0.0594838
[22]	training's auc: 0.802066	training's binary_logloss: 0.0875477	valid_1's auc: 0.810087	valid_1's binary_logloss: 0.0592732
[23]	training's auc: 0.802621	training's binary_logloss: 0.087402	valid_1's auc: 0.809696	valid_1's binary_logloss: 0.0591517
[24]	training's auc: 0.802456	training's binary_logloss: 0.0872056	valid_1's auc: 0.809599	valid_1's binary_logloss: 0.0589684
[25]	training's auc: 0.802714	training's binary_logloss: 0.0870772	valid_1's auc: 0.809573	valid_1's binary_logloss: 0.0588481
[26]	training's auc: 0.803158	training's binary_logloss: 0.0868977	valid_1's auc: 0.809865	valid_1's binary_logloss: 0.0586947
[27]	training's auc: 0.803217	training's binary_logloss: 0.086744	valid_1's auc: 0.809628	valid_1's binary_logloss: 0.0585383
[28]	training's auc: 0.804252	training's binary_logloss: 0.0866573	valid_1's auc: 0.811128	valid_1's binary_logloss: 0.0584608
[29]	training's auc: 0.804553	training's binary_logloss: 0.0865065	valid_1's auc: 0.811729	valid_1's binary_logloss: 0.0583086
[30]	training's auc: 0.804338	training's binary_logloss: 0.0863911	valid_1's auc: 0.811911	valid_1's binary_logloss: 0.0582117
[31]	training's auc: 0.804622	training's binary_logloss: 0.0862567	valid_1's auc: 0.811873	valid_1's binary_logloss: 0.0580706
[32]	training's auc: 0.80478	training's binary_logloss: 0.0861849	valid_1's auc: 0.811709	valid_1's binary_logloss: 0.0580063
[33]	training's auc: 0.805155	training's binary_logloss: 0.0860736	valid_1's auc: 0.811696	valid_1's binary_logloss: 0.0578887
[34]	training's auc: 0.804984	training's binary_logloss: 0.0860161	valid_1's auc: 0.811514	valid_1's binary_logloss: 0.0578275
[35]	training's auc: 0.805	training's binary_logloss: 0.0859207	valid_1's auc: 0.811745	valid_1's binary_logloss: 0.0577335
[36]	training's auc: 0.805193	training's binary_logloss: 0.0858218	valid_1's auc: 0.812049	valid_1's binary_logloss: 0.0576406
[37]	training's auc: 0.805235	training's binary_logloss: 0.0857342	valid_1's auc: 0.812539	valid_1's binary_logloss: 0.0575412
[38]	training's auc: 0.805632	training's binary_logloss: 0.0856421	valid_1's auc: 0.81291	valid_1's binary_logloss: 0.0574406
[39]	training's auc: 0.805991	training's binary_logloss: 0.0855961	valid_1's auc: 0.813657	valid_1's binary_logloss: 0.0574029
[40]	training's auc: 0.806425	training's binary_logloss: 0.0855149	valid_1's auc: 0.814025	valid_1's binary_logloss: 0.0573064
[41]	training's auc: 0.806606	training's binary_logloss: 0.0854352	valid_1's auc: 0.814374	valid_1's binary_logloss: 0.0572174
[42]	training's auc: 0.806573	training's binary_logloss: 0.0853742	valid_1's auc: 0.814086	valid_1's binary_logloss: 0.0571594
[43]	training's auc: 0.806743	training's binary_logloss: 0.0853023	valid_1's auc: 0.814595	valid_1's binary_logloss: 0.0570839
[44]	training's auc: 0.806743	training's binary_logloss: 0.0852461	valid_1's auc: 0.815221	valid_1's binary_logloss: 0.0570194
[45]	training's auc: 0.806815	training's binary_logloss: 0.0851931	valid_1's auc: 0.81454	valid_1's binary_logloss: 0.0569637
[46]	training's auc: 0.807021	training's binary_logloss: 0.085143	valid_1's auc: 0.814218	valid_1's binary_logloss: 0.0569145
[47]	training's auc: 0.806955	training's binary_logloss: 0.0850872	valid_1's auc: 0.814365	valid_1's binary_logloss: 0.0568348
[48]	training's auc: 0.807045	training's binary_logloss: 0.0850439	valid_1's auc: 0.814749	valid_1's binary_logloss: 0.0567858
[49]	training's auc: 0.807163	training's binary_logloss: 0.0850047	valid_1's auc: 0.815054	valid_1's binary_logloss: 0.0567532
[50]	training's auc: 0.807072	training's binary_logloss: 0.0849657	valid_1's auc: 0.81448	valid_1's binary_logloss: 0.0567092
[51]	training's auc: 0.807059	training's binary_logloss: 0.0849375	valid_1's auc: 0.814442	valid_1's binary_logloss: 0.056675
[52]	training's auc: 0.807123	training's binary_logloss: 0.0848976	valid_1's auc: 0.814686	valid_1's binary_logloss: 0.0566136
[53]	training's auc: 0.807273	training's binary_logloss: 0.0848829	valid_1's auc: 0.815037	valid_1's binary_logloss: 0.0566088
[54]	training's auc: 0.807267	training's binary_logloss: 0.0848495	valid_1's auc: 0.814929	valid_1's binary_logloss: 0.0565668
[55]	training's auc: 0.807479	training's binary_logloss: 0.0848183	valid_1's auc: 0.815321	valid_1's binary_logloss: 0.056538
[56]	training's auc: 0.807608	training's binary_logloss: 0.0847882	valid_1's auc: 0.815785	valid_1's binary_logloss: 0.0565081
[57]	training's auc: 0.807705	training's binary_logloss: 0.0847571	valid_1's auc: 0.81586	valid_1's binary_logloss: 0.0564711
[58]	training's auc: 0.807829	training's binary_logloss: 0.0847242	valid_1's auc: 0.816078	valid_1's binary_logloss: 0.0564347
[59]	training's auc: 0.807836	training's binary_logloss: 0.0846966	valid_1's auc: 0.815969	valid_1's binary_logloss: 0.0564051
[60]	training's auc: 0.807932	training's binary_logloss: 0.084677	valid_1's auc: 0.816165	valid_1's binary_logloss: 0.0563777
[61]	training's auc: 0.808128	training's binary_logloss: 0.0846542	valid_1's auc: 0.816272	valid_1's binary_logloss: 0.0563616
[62]	training's auc: 0.808032	training's binary_logloss: 0.084626	valid_1's auc: 0.816418	valid_1's binary_logloss: 0.0563228
[63]	training's auc: 0.808202	training's binary_logloss: 0.0846007	valid_1's auc: 0.816501	valid_1's binary_logloss: 0.0562933
[64]	training's auc: 0.808293	training's binary_logloss: 0.0845614	valid_1's auc: 0.816811	valid_1's binary_logloss: 0.0562547
[65]	training's auc: 0.808388	training's binary_logloss: 0.0845395	valid_1's auc: 0.81681	valid_1's binary_logloss: 0.0562309
[66]	training's auc: 0.808498	training's binary_logloss: 0.0845224	valid_1's auc: 0.81661	valid_1's binary_logloss: 0.0562206
[67]	training's auc: 0.808623	training's binary_logloss: 0.0845026	valid_1's auc: 0.816483	valid_1's binary_logloss: 0.0561998
[68]	training's auc: 0.808581	training's binary_logloss: 0.0844806	valid_1's auc: 0.816755	valid_1's binary_logloss: 0.0561771
[69]	training's auc: 0.808788	training's binary_logloss: 0.0844683	valid_1's auc: 0.81722	valid_1's binary_logloss: 0.0561672
[70]	training's auc: 0.808805	training's binary_logloss: 0.0844468	valid_1's auc: 0.817082	valid_1's binary_logloss: 0.0561362
[71]	training's auc: 0.809022	training's binary_logloss: 0.084432	valid_1's auc: 0.817074	valid_1's binary_logloss: 0.0561312
[72]	training's auc: 0.809063	training's binary_logloss: 0.0844117	valid_1's auc: 0.816864	valid_1's binary_logloss: 0.0561104
[73]	training's auc: 0.809107	training's binary_logloss: 0.0843966	valid_1's auc: 0.816877	valid_1's binary_logloss: 0.0560993
[74]	training's auc: 0.809203	training's binary_logloss: 0.0843875	valid_1's auc: 0.816657	valid_1's binary_logloss: 0.0560922
[75]	training's auc: 0.809259	training's binary_logloss: 0.0843759	valid_1's auc: 0.816402	valid_1's binary_logloss: 0.056085
[76]	training's auc: 0.809361	training's binary_logloss: 0.0843586	valid_1's auc: 0.816517	valid_1's binary_logloss: 0.0560751
[77]	training's auc: 0.809421	training's binary_logloss: 0.084339	valid_1's auc: 0.816594	valid_1's binary_logloss: 0.0560546
[78]	training's auc: 0.809535	training's binary_logloss: 0.0843318	valid_1's auc: 0.816616	valid_1's binary_logloss: 0.0560484
[79]	training's auc: 0.80968	training's binary_logloss: 0.0843194	valid_1's auc: 0.816911	valid_1's binary_logloss: 0.0560291
[80]	training's auc: 0.8098	training's binary_logloss: 0.0843084	valid_1's auc: 0.816888	valid_1's binary_logloss: 0.0560224
[81]	training's auc: 0.809876	training's binary_logloss: 0.0842951	valid_1's auc: 0.816484	valid_1's binary_logloss: 0.0560072
[82]	training's auc: 0.809964	training's binary_logloss: 0.0842853	valid_1's auc: 0.816303	valid_1's binary_logloss: 0.0560108
[83]	training's auc: 0.810191	training's binary_logloss: 0.0842674	valid_1's auc: 0.816584	valid_1's binary_logloss: 0.0559851
[84]	training's auc: 0.810264	training's binary_logloss: 0.0842563	valid_1's auc: 0.816697	valid_1's binary_logloss: 0.0559769
[85]	training's auc: 0.810346	training's binary_logloss: 0.0842448	valid_1's auc: 0.816915	valid_1's binary_logloss: 0.0559724
[86]	training's auc: 0.81037	training's binary_logloss: 0.084233	valid_1's auc: 0.816981	valid_1's binary_logloss: 0.0559618
[87]	training's auc: 0.810337	training's binary_logloss: 0.0842191	valid_1's auc: 0.816701	valid_1's binary_logloss: 0.0559556
[88]	training's auc: 0.81048	training's binary_logloss: 0.0842017	valid_1's auc: 0.816641	valid_1's binary_logloss: 0.0559454
[89]	training's auc: 0.810608	training's binary_logloss: 0.0841822	valid_1's auc: 0.816591	valid_1's binary_logloss: 0.0559269
[90]	training's auc: 0.810655	training's binary_logloss: 0.0841739	valid_1's auc: 0.816726	valid_1's binary_logloss: 0.0559173
[91]	training's auc: 0.810719	training's binary_logloss: 0.0841643	valid_1's auc: 0.816838	valid_1's binary_logloss: 0.0559096
[92]	training's auc: 0.810818	training's binary_logloss: 0.084155	valid_1's auc: 0.817027	valid_1's binary_logloss: 0.0558909
[93]	training's auc: 0.810814	training's binary_logloss: 0.0841443	valid_1's auc: 0.81709	valid_1's binary_logloss: 0.0558779
[94]	training's auc: 0.810849	training's binary_logloss: 0.0841328	valid_1's auc: 0.816985	valid_1's binary_logloss: 0.0558654
[95]	training's auc: 0.810878	training's binary_logloss: 0.0841239	valid_1's auc: 0.817107	valid_1's binary_logloss: 0.0558463
[96]	training's auc: 0.810903	training's binary_logloss: 0.0841197	valid_1's auc: 0.816735	valid_1's binary_logloss: 0.0558617
[97]	training's auc: 0.810934	training's binary_logloss: 0.0841127	valid_1's auc: 0.816641	valid_1's binary_logloss: 0.0558565
[98]	training's auc: 0.810958	training's binary_logloss: 0.0841034	valid_1's auc: 0.816856	valid_1's binary_logloss: 0.0558393
[99]	training's auc: 0.811096	training's binary_logloss: 0.0840931	valid_1's auc: 0.816683	valid_1's binary_logloss: 0.0558354
[100]	training's auc: 0.811127	training's binary_logloss: 0.0840849	valid_1's auc: 0.816753	valid_1's binary_logloss: 0.0558261
[101]	training's auc: 0.811209	training's binary_logloss: 0.084073	valid_1's auc: 0.81679	valid_1's binary_logloss: 0.0558183
[102]	training's auc: 0.811326	training's binary_logloss: 0.0840615	valid_1's auc: 0.816839	valid_1's binary_logloss: 0.0558055
[103]	training's auc: 0.811331	training's binary_logloss: 0.084058	valid_1's auc: 0.816707	valid_1's binary_logloss: 0.055805
[104]	training's auc: 0.811495	training's binary_logloss: 0.0840347	valid_1's auc: 0.816186	valid_1's binary_logloss: 0.0558033
[105]	training's auc: 0.811504	training's binary_logloss: 0.0840285	valid_1's auc: 0.815966	valid_1's binary_logloss: 0.0558024
[106]	training's auc: 0.811606	training's binary_logloss: 0.0840172	valid_1's auc: 0.815874	valid_1's binary_logloss: 0.0557894
[107]	training's auc: 0.811649	training's binary_logloss: 0.0840077	valid_1's auc: 0.816063	valid_1's binary_logloss: 0.0557843
[108]	training's auc: 0.811723	training's binary_logloss: 0.0840034	valid_1's auc: 0.815978	valid_1's binary_logloss: 0.0557805
[109]	training's auc: 0.811758	training's binary_logloss: 0.0839971	valid_1's auc: 0.816078	valid_1's binary_logloss: 0.0557829
[110]	training's auc: 0.811774	training's binary_logloss: 0.0839932	valid_1's auc: 0.816043	valid_1's binary_logloss: 0.0557805
[111]	training's auc: 0.811827	training's binary_logloss: 0.0839894	valid_1's auc: 0.816089	valid_1's binary_logloss: 0.0557754
[112]	training's auc: 0.811903	training's binary_logloss: 0.0839828	valid_1's auc: 0.816169	valid_1's binary_logloss: 0.0557696
[113]	training's auc: 0.811885	training's binary_logloss: 0.0839783	valid_1's auc: 0.81611	valid_1's binary_logloss: 0.055759
[114]	training's auc: 0.812008	training's binary_logloss: 0.0839713	valid_1's auc: 0.816059	valid_1's binary_logloss: 0.0557656
[115]	training's auc: 0.812011	training's binary_logloss: 0.0839668	valid_1's auc: 0.81613	valid_1's binary_logloss: 0.0557569
[116]	training's auc: 0.812099	training's binary_logloss: 0.0839602	valid_1's auc: 0.816182	valid_1's binary_logloss: 0.0557503
[117]	training's auc: 0.812161	training's binary_logloss: 0.0839484	valid_1's auc: 0.816163	valid_1's binary_logloss: 0.0557415
[118]	training's auc: 0.812206	training's binary_logloss: 0.083943	valid_1's auc: 0.81609	valid_1's binary_logloss: 0.055739
[119]	training's auc: 0.812265	training's binary_logloss: 0.0839348	valid_1's auc: 0.816047	valid_1's binary_logloss: 0.0557407
[120]	training's auc: 0.812401	training's binary_logloss: 0.0839245	valid_1's auc: 0.816042	valid_1's binary_logloss: 0.0557514
[121]	training's auc: 0.812466	training's binary_logloss: 0.0839183	valid_1's auc: 0.816251	valid_1's binary_logloss: 0.0557414
[122]	training's auc: 0.812562	training's binary_logloss: 0.0839112	valid_1's auc: 0.816465	valid_1's binary_logloss: 0.055734
[123]	training's auc: 0.812579	training's binary_logloss: 0.0839054	valid_1's auc: 0.816412	valid_1's binary_logloss: 0.0557337
[124]	training's auc: 0.812694	training's binary_logloss: 0.0838958	valid_1's auc: 0.816396	valid_1's binary_logloss: 0.0557388
[125]	training's auc: 0.812792	training's binary_logloss: 0.0838834	valid_1's auc: 0.816347	valid_1's binary_logloss: 0.055728
[126]	training's auc: 0.812853	training's binary_logloss: 0.0838757	valid_1's auc: 0.816155	valid_1's binary_logloss: 0.0557399
[127]	training's auc: 0.812859	training's binary_logloss: 0.0838717	valid_1's auc: 0.816179	valid_1's binary_logloss: 0.0557382
[128]	training's auc: 0.81286	training's binary_logloss: 0.0838678	valid_1's auc: 0.816188	valid_1's binary_logloss: 0.0557338
[129]	training's auc: 0.812902	training's binary_logloss: 0.083861	valid_1's auc: 0.816339	valid_1's binary_logloss: 0.0557306
[130]	training's auc: 0.812934	training's binary_logloss: 0.0838547	valid_1's auc: 0.816338	valid_1's binary_logloss: 0.0557402
[131]	training's auc: 0.812968	training's binary_logloss: 0.0838511	valid_1's auc: 0.816398	valid_1's binary_logloss: 0.0557361
[132]	training's auc: 0.812989	training's binary_logloss: 0.0838471	valid_1's auc: 0.816365	valid_1's binary_logloss: 0.0557305
[133]	training's auc: 0.813074	training's binary_logloss: 0.0838381	valid_1's auc: 0.816402	valid_1's binary_logloss: 0.055732
[134]	training's auc: 0.813081	training's binary_logloss: 0.083832	valid_1's auc: 0.816407	valid_1's binary_logloss: 0.0557204
[135]	training's auc: 0.813148	training's binary_logloss: 0.0838236	valid_1's auc: 0.816402	valid_1's binary_logloss: 0.0557181
[136]	training's auc: 0.813278	training's binary_logloss: 0.0838136	valid_1's auc: 0.816311	valid_1's binary_logloss: 0.0557265
[137]	training's auc: 0.813272	training's binary_logloss: 0.0838089	valid_1's auc: 0.816138	valid_1's binary_logloss: 0.0557348
[138]	training's auc: 0.813294	training's binary_logloss: 0.0838043	valid_1's auc: 0.816023	valid_1's binary_logloss: 0.0557341
[139]	training's auc: 0.813322	training's binary_logloss: 0.083801	valid_1's auc: 0.815985	valid_1's binary_logloss: 0.0557357
[140]	training's auc: 0.813366	training's binary_logloss: 0.0837942	valid_1's auc: 0.815988	valid_1's binary_logloss: 0.0557344
[141]	training's auc: 0.813386	training's binary_logloss: 0.0837917	valid_1's auc: 0.816145	valid_1's binary_logloss: 0.0557285
[142]	training's auc: 0.81343	training's binary_logloss: 0.0837881	valid_1's auc: 0.816108	valid_1's binary_logloss: 0.0557306
[143]	training's auc: 0.813439	training's binary_logloss: 0.0837797	valid_1's auc: 0.816183	valid_1's binary_logloss: 0.0557132
[144]	training's auc: 0.813438	training's binary_logloss: 0.0837708	valid_1's auc: 0.815809	valid_1's binary_logloss: 0.0557276
[145]	training's auc: 0.813518	training's binary_logloss: 0.0837577	valid_1's auc: 0.815753	valid_1's binary_logloss: 0.055723
[146]	training's auc: 0.813544	training's binary_logloss: 0.0837528	valid_1's auc: 0.81582	valid_1's binary_logloss: 0.0557235
[147]	training's auc: 0.813547	training's binary_logloss: 0.0837494	valid_1's auc: 0.815739	valid_1's binary_logloss: 0.0557281
[148]	training's auc: 0.813601	training's binary_logloss: 0.0837412	valid_1's auc: 0.815785	valid_1's binary_logloss: 0.0557239
[149]	training's auc: 0.813631	training's binary_logloss: 0.0837374	valid_1's auc: 0.815803	valid_1's binary_logloss: 0.0557217
[150]	training's auc: 0.813702	training's binary_logloss: 0.0837295	valid_1's auc: 0.815734	valid_1's binary_logloss: 0.0557281
[151]	training's auc: 0.81378	training's binary_logloss: 0.0837239	valid_1's auc: 0.8157	valid_1's binary_logloss: 0.0557294
[152]	training's auc: 0.813797	training's binary_logloss: 0.0837201	valid_1's auc: 0.815536	valid_1's binary_logloss: 0.0557362
[153]	training's auc: 0.813909	training's binary_logloss: 0.0837115	valid_1's auc: 0.815413	valid_1's binary_logloss: 0.0557369
[154]	training's auc: 0.814048	training's binary_logloss: 0.083701	valid_1's auc: 0.815291	valid_1's binary_logloss: 0.0557463
[155]	training's auc: 0.814129	training's binary_logloss: 0.0836938	valid_1's auc: 0.815297	valid_1's binary_logloss: 0.0557441
[156]	training's auc: 0.81416	training's binary_logloss: 0.0836791	valid_1's auc: 0.81537	valid_1's binary_logloss: 0.0557294
[157]	training's auc: 0.814207	training's binary_logloss: 0.0836744	valid_1's auc: 0.815386	valid_1's binary_logloss: 0.0557321
[158]	training's auc: 0.814229	training's binary_logloss: 0.0836614	valid_1's auc: 0.815496	valid_1's binary_logloss: 0.0557198
[159]	training's auc: 0.814273	training's binary_logloss: 0.0836503	valid_1's auc: 0.815553	valid_1's binary_logloss: 0.0557219
[160]	training's auc: 0.814337	training's binary_logloss: 0.0836455	valid_1's auc: 0.815676	valid_1's binary_logloss: 0.0557211
[161]	training's auc: 0.814365	training's binary_logloss: 0.083643	valid_1's auc: 0.815727	valid_1's binary_logloss: 0.0557114
[162]	training's auc: 0.814391	training's binary_logloss: 0.0836374	valid_1's auc: 0.81575	valid_1's binary_logloss: 0.0557144
[163]	training's auc: 0.814438	training's binary_logloss: 0.0836316	valid_1's auc: 0.815773	valid_1's binary_logloss: 0.0557121
[164]	training's auc: 0.814479	training's binary_logloss: 0.0836259	valid_1's auc: 0.815514	valid_1's binary_logloss: 0.0557137
[165]	training's auc: 0.814536	training's binary_logloss: 0.0836069	valid_1's auc: 0.815567	valid_1's binary_logloss: 0.0557065
[166]	training's auc: 0.81458	training's binary_logloss: 0.0836019	valid_1's auc: 0.81561	valid_1's binary_logloss: 0.0557044
[167]	training's auc: 0.814629	training's binary_logloss: 0.0835964	valid_1's auc: 0.815543	valid_1's binary_logloss: 0.0557057
[168]	training's auc: 0.814677	training's binary_logloss: 0.0835905	valid_1's auc: 0.815461	valid_1's binary_logloss: 0.05571
[169]	training's auc: 0.814701	training's binary_logloss: 0.0835872	valid_1's auc: 0.815485	valid_1's binary_logloss: 0.0557125
Early stopping, best iteration is:
[69]	training's auc: 0.808788	training's binary_logloss: 0.0844683	valid_1's auc: 0.81722	valid_1's binary_logloss: 0.0561672
8
[1]	training's auc: 0.651876	training's binary_logloss: 0.0966336	valid_1's auc: 0.672491	valid_1's binary_logloss: 0.0712814
Training until validation scores don't improve for 100 rounds.
[2]	training's auc: 0.782222	training's binary_logloss: 0.0952334	valid_1's auc: 0.796897	valid_1's binary_logloss: 0.0703147
[3]	training's auc: 0.782296	training's binary_logloss: 0.0946264	valid_1's auc: 0.797943	valid_1's binary_logloss: 0.0698724
[4]	training's auc: 0.78281	training's binary_logloss: 0.0940255	valid_1's auc: 0.796878	valid_1's binary_logloss: 0.0693916
[5]	training's auc: 0.790877	training's binary_logloss: 0.0930066	valid_1's auc: 0.802175	valid_1's binary_logloss: 0.0687081
[6]	training's auc: 0.795996	training's binary_logloss: 0.0920978	valid_1's auc: 0.812757	valid_1's binary_logloss: 0.068065
[7]	training's auc: 0.794912	training's binary_logloss: 0.0916311	valid_1's auc: 0.810214	valid_1's binary_logloss: 0.0676846
[8]	training's auc: 0.795266	training's binary_logloss: 0.0909168	valid_1's auc: 0.808938	valid_1's binary_logloss: 0.0671766
[9]	training's auc: 0.798298	training's binary_logloss: 0.0907412	valid_1's auc: 0.815323	valid_1's binary_logloss: 0.0669682
[10]	training's auc: 0.797394	training's binary_logloss: 0.0904158	valid_1's auc: 0.814292	valid_1's binary_logloss: 0.0666602
[11]	training's auc: 0.797961	training's binary_logloss: 0.0897895	valid_1's auc: 0.813742	valid_1's binary_logloss: 0.0662159
[12]	training's auc: 0.799646	training's binary_logloss: 0.089227	valid_1's auc: 0.817797	valid_1's binary_logloss: 0.065767
[13]	training's auc: 0.798912	training's binary_logloss: 0.0889946	valid_1's auc: 0.817083	valid_1's binary_logloss: 0.0655623
[14]	training's auc: 0.798524	training's binary_logloss: 0.0887722	valid_1's auc: 0.816027	valid_1's binary_logloss: 0.065345
[15]	training's auc: 0.799311	training's binary_logloss: 0.0883419	valid_1's auc: 0.816868	valid_1's binary_logloss: 0.0650306
[16]	training's auc: 0.800168	training's binary_logloss: 0.0879358	valid_1's auc: 0.818169	valid_1's binary_logloss: 0.064701
[17]	training's auc: 0.800163	training's binary_logloss: 0.0877144	valid_1's auc: 0.817945	valid_1's binary_logloss: 0.0644967
[18]	training's auc: 0.800814	training's binary_logloss: 0.0873888	valid_1's auc: 0.819244	valid_1's binary_logloss: 0.0642213
[19]	training's auc: 0.800638	training's binary_logloss: 0.08709	valid_1's auc: 0.819727	valid_1's binary_logloss: 0.0639646
[20]	training's auc: 0.800659	training's binary_logloss: 0.0869065	valid_1's auc: 0.819501	valid_1's binary_logloss: 0.0637845
[21]	training's auc: 0.801281	training's binary_logloss: 0.0866696	valid_1's auc: 0.819769	valid_1's binary_logloss: 0.0635872
[22]	training's auc: 0.801517	training's binary_logloss: 0.0864386	valid_1's auc: 0.820189	valid_1's binary_logloss: 0.0633721
[23]	training's auc: 0.802045	training's binary_logloss: 0.0862893	valid_1's auc: 0.820194	valid_1's binary_logloss: 0.0632338
[24]	training's auc: 0.802062	training's binary_logloss: 0.0860903	valid_1's auc: 0.820576	valid_1's binary_logloss: 0.0630546
[25]	training's auc: 0.802013	training's binary_logloss: 0.0859674	valid_1's auc: 0.82046	valid_1's binary_logloss: 0.0629284
[26]	training's auc: 0.802355	training's binary_logloss: 0.0858004	valid_1's auc: 0.821287	valid_1's binary_logloss: 0.0627763
[27]	training's auc: 0.802812	training's binary_logloss: 0.0856565	valid_1's auc: 0.821568	valid_1's binary_logloss: 0.0626215
[28]	training's auc: 0.803651	training's binary_logloss: 0.0855865	valid_1's auc: 0.820862	valid_1's binary_logloss: 0.0625611
[29]	training's auc: 0.803934	training's binary_logloss: 0.0854361	valid_1's auc: 0.821599	valid_1's binary_logloss: 0.0624167
[30]	training's auc: 0.804023	training's binary_logloss: 0.0853239	valid_1's auc: 0.821448	valid_1's binary_logloss: 0.0623072
[31]	training's auc: 0.804164	training's binary_logloss: 0.0851905	valid_1's auc: 0.822445	valid_1's binary_logloss: 0.0621656
[32]	training's auc: 0.804365	training's binary_logloss: 0.0851225	valid_1's auc: 0.822411	valid_1's binary_logloss: 0.0620876
[33]	training's auc: 0.804282	training's binary_logloss: 0.0850101	valid_1's auc: 0.822785	valid_1's binary_logloss: 0.0619808
[34]	training's auc: 0.804337	training's binary_logloss: 0.0849559	valid_1's auc: 0.823679	valid_1's binary_logloss: 0.0619049
[35]	training's auc: 0.804402	training's binary_logloss: 0.0848642	valid_1's auc: 0.824117	valid_1's binary_logloss: 0.0618065
[36]	training's auc: 0.804437	training's binary_logloss: 0.0847682	valid_1's auc: 0.82468	valid_1's binary_logloss: 0.0616965
[37]	training's auc: 0.804705	training's binary_logloss: 0.0846795	valid_1's auc: 0.824886	valid_1's binary_logloss: 0.0616042
[38]	training's auc: 0.804914	training's binary_logloss: 0.0845808	valid_1's auc: 0.824717	valid_1's binary_logloss: 0.0615223
[39]	training's auc: 0.80542	training's binary_logloss: 0.0845329	valid_1's auc: 0.824429	valid_1's binary_logloss: 0.0614744
[40]	training's auc: 0.80561	training's binary_logloss: 0.0844504	valid_1's auc: 0.823931	valid_1's binary_logloss: 0.0613888
[41]	training's auc: 0.805765	training's binary_logloss: 0.0843685	valid_1's auc: 0.823855	valid_1's binary_logloss: 0.0613149
[42]	training's auc: 0.805723	training's binary_logloss: 0.084313	valid_1's auc: 0.824308	valid_1's binary_logloss: 0.0612435
[43]	training's auc: 0.805933	training's binary_logloss: 0.0842366	valid_1's auc: 0.824345	valid_1's binary_logloss: 0.0611773
[44]	training's auc: 0.805917	training's binary_logloss: 0.0841822	valid_1's auc: 0.82491	valid_1's binary_logloss: 0.0611007
[45]	training's auc: 0.806125	training's binary_logloss: 0.0841292	valid_1's auc: 0.82475	valid_1's binary_logloss: 0.0610397
[46]	training's auc: 0.806393	training's binary_logloss: 0.0840789	valid_1's auc: 0.824461	valid_1's binary_logloss: 0.0609927
[47]	training's auc: 0.806412	training's binary_logloss: 0.0840263	valid_1's auc: 0.824824	valid_1's binary_logloss: 0.0609207
[48]	training's auc: 0.80666	training's binary_logloss: 0.0839782	valid_1's auc: 0.825245	valid_1's binary_logloss: 0.0608693
[49]	training's auc: 0.806707	training's binary_logloss: 0.0839407	valid_1's auc: 0.825594	valid_1's binary_logloss: 0.0608292
[50]	training's auc: 0.80672	training's binary_logloss: 0.0839073	valid_1's auc: 0.825844	valid_1's binary_logloss: 0.0607807
[51]	training's auc: 0.806756	training's binary_logloss: 0.0838608	valid_1's auc: 0.825818	valid_1's binary_logloss: 0.060727
[52]	training's auc: 0.806669	training's binary_logloss: 0.0838224	valid_1's auc: 0.82552	valid_1's binary_logloss: 0.06067
[53]	training's auc: 0.806651	training's binary_logloss: 0.0838034	valid_1's auc: 0.825846	valid_1's binary_logloss: 0.060648
[54]	training's auc: 0.806617	training's binary_logloss: 0.0837733	valid_1's auc: 0.825788	valid_1's binary_logloss: 0.0606121
[55]	training's auc: 0.806792	training's binary_logloss: 0.0837419	valid_1's auc: 0.825677	valid_1's binary_logloss: 0.0605781
[56]	training's auc: 0.807028	training's binary_logloss: 0.0837113	valid_1's auc: 0.825301	valid_1's binary_logloss: 0.0605593
[57]	training's auc: 0.807191	training's binary_logloss: 0.0836711	valid_1's auc: 0.825041	valid_1's binary_logloss: 0.0605329
[58]	training's auc: 0.807456	training's binary_logloss: 0.0836309	valid_1's auc: 0.824855	valid_1's binary_logloss: 0.0605086
[59]	training's auc: 0.807649	training's binary_logloss: 0.0836081	valid_1's auc: 0.824865	valid_1's binary_logloss: 0.0604993
[60]	training's auc: 0.807782	training's binary_logloss: 0.083587	valid_1's auc: 0.824816	valid_1's binary_logloss: 0.0604744
[61]	training's auc: 0.807984	training's binary_logloss: 0.0835607	valid_1's auc: 0.824744	valid_1's binary_logloss: 0.0604517
[62]	training's auc: 0.807988	training's binary_logloss: 0.0835311	valid_1's auc: 0.824757	valid_1's binary_logloss: 0.0604051
[63]	training's auc: 0.808153	training's binary_logloss: 0.0835002	valid_1's auc: 0.824989	valid_1's binary_logloss: 0.0603817
[64]	training's auc: 0.808191	training's binary_logloss: 0.0834812	valid_1's auc: 0.824957	valid_1's binary_logloss: 0.0603725
[65]	training's auc: 0.808286	training's binary_logloss: 0.0834613	valid_1's auc: 0.824772	valid_1's binary_logloss: 0.0603514
[66]	training's auc: 0.808291	training's binary_logloss: 0.0834439	valid_1's auc: 0.824602	valid_1's binary_logloss: 0.0603396
[67]	training's auc: 0.808445	training's binary_logloss: 0.0834233	valid_1's auc: 0.824643	valid_1's binary_logloss: 0.0603051
[68]	training's auc: 0.808687	training's binary_logloss: 0.0833945	valid_1's auc: 0.824499	valid_1's binary_logloss: 0.0603005
[69]	training's auc: 0.808858	training's binary_logloss: 0.0833786	valid_1's auc: 0.824734	valid_1's binary_logloss: 0.0602759
[70]	training's auc: 0.808975	training's binary_logloss: 0.0833647	valid_1's auc: 0.824802	valid_1's binary_logloss: 0.0602611
[71]	training's auc: 0.809138	training's binary_logloss: 0.0833506	valid_1's auc: 0.824615	valid_1's binary_logloss: 0.0602555
[72]	training's auc: 0.80915	training's binary_logloss: 0.0833292	valid_1's auc: 0.824555	valid_1's binary_logloss: 0.0602288
[73]	training's auc: 0.809218	training's binary_logloss: 0.083313	valid_1's auc: 0.824424	valid_1's binary_logloss: 0.0602239
[74]	training's auc: 0.809292	training's binary_logloss: 0.0833036	valid_1's auc: 0.824408	valid_1's binary_logloss: 0.0602215
[75]	training's auc: 0.809399	training's binary_logloss: 0.083288	valid_1's auc: 0.824159	valid_1's binary_logloss: 0.0602156
[76]	training's auc: 0.809528	training's binary_logloss: 0.0832685	valid_1's auc: 0.824594	valid_1's binary_logloss: 0.0601978
[77]	training's auc: 0.809609	training's binary_logloss: 0.0832551	valid_1's auc: 0.824379	valid_1's binary_logloss: 0.0601937
[78]	training's auc: 0.809667	training's binary_logloss: 0.0832479	valid_1's auc: 0.824251	valid_1's binary_logloss: 0.0602006
[79]	training's auc: 0.809771	training's binary_logloss: 0.0832364	valid_1's auc: 0.82418	valid_1's binary_logloss: 0.0601822
[80]	training's auc: 0.809818	training's binary_logloss: 0.0832251	valid_1's auc: 0.823997	valid_1's binary_logloss: 0.0601747
[81]	training's auc: 0.809845	training's binary_logloss: 0.0831997	valid_1's auc: 0.823921	valid_1's binary_logloss: 0.0601597
[82]	training's auc: 0.809981	training's binary_logloss: 0.0831874	valid_1's auc: 0.823746	valid_1's binary_logloss: 0.0601594
[83]	training's auc: 0.810113	training's binary_logloss: 0.0831691	valid_1's auc: 0.823861	valid_1's binary_logloss: 0.0601302
[84]	training's auc: 0.810134	training's binary_logloss: 0.08316	valid_1's auc: 0.823783	valid_1's binary_logloss: 0.0601252
[85]	training's auc: 0.810189	training's binary_logloss: 0.0831473	valid_1's auc: 0.823775	valid_1's binary_logloss: 0.0601218
[86]	training's auc: 0.810156	training's binary_logloss: 0.0831307	valid_1's auc: 0.82359	valid_1's binary_logloss: 0.0601234
[87]	training's auc: 0.810128	training's binary_logloss: 0.0831154	valid_1's auc: 0.823591	valid_1's binary_logloss: 0.0601117
[88]	training's auc: 0.810232	training's binary_logloss: 0.083101	valid_1's auc: 0.823949	valid_1's binary_logloss: 0.0600746
[89]	training's auc: 0.810245	training's binary_logloss: 0.0830823	valid_1's auc: 0.823989	valid_1's binary_logloss: 0.0600623
[90]	training's auc: 0.810291	training's binary_logloss: 0.0830743	valid_1's auc: 0.823931	valid_1's binary_logloss: 0.0600605
[91]	training's auc: 0.810381	training's binary_logloss: 0.0830613	valid_1's auc: 0.823988	valid_1's binary_logloss: 0.060055
[92]	training's auc: 0.810537	training's binary_logloss: 0.0830497	valid_1's auc: 0.823938	valid_1's binary_logloss: 0.0600608
[93]	training's auc: 0.810596	training's binary_logloss: 0.0830327	valid_1's auc: 0.823778	valid_1's binary_logloss: 0.0600516
[94]	training's auc: 0.810681	training's binary_logloss: 0.0830257	valid_1's auc: 0.823809	valid_1's binary_logloss: 0.0600451
[95]	training's auc: 0.810708	training's binary_logloss: 0.0830142	valid_1's auc: 0.82377	valid_1's binary_logloss: 0.0600337
[96]	training's auc: 0.81068	training's binary_logloss: 0.083004	valid_1's auc: 0.823705	valid_1's binary_logloss: 0.0600304
[97]	training's auc: 0.810811	training's binary_logloss: 0.0829835	valid_1's auc: 0.823786	valid_1's binary_logloss: 0.0600237
[98]	training's auc: 0.810819	training's binary_logloss: 0.0829762	valid_1's auc: 0.823652	valid_1's binary_logloss: 0.0600215
[99]	training's auc: 0.810861	training's binary_logloss: 0.0829698	valid_1's auc: 0.823793	valid_1's binary_logloss: 0.0600055
[100]	training's auc: 0.810882	training's binary_logloss: 0.0829642	valid_1's auc: 0.823639	valid_1's binary_logloss: 0.0600004
[101]	training's auc: 0.810986	training's binary_logloss: 0.0829498	valid_1's auc: 0.823674	valid_1's binary_logloss: 0.0599905
[102]	training's auc: 0.811097	training's binary_logloss: 0.0829382	valid_1's auc: 0.823761	valid_1's binary_logloss: 0.0599795
[103]	training's auc: 0.811146	training's binary_logloss: 0.0829249	valid_1's auc: 0.823688	valid_1's binary_logloss: 0.0599688
[104]	training's auc: 0.81129	training's binary_logloss: 0.0829014	valid_1's auc: 0.823731	valid_1's binary_logloss: 0.059965
[105]	training's auc: 0.81128	training's binary_logloss: 0.0828934	valid_1's auc: 0.823728	valid_1's binary_logloss: 0.0599626
[106]	training's auc: 0.811379	training's binary_logloss: 0.0828801	valid_1's auc: 0.823626	valid_1's binary_logloss: 0.0599544
[107]	training's auc: 0.811452	training's binary_logloss: 0.08287	valid_1's auc: 0.82364	valid_1's binary_logloss: 0.0599516
[108]	training's auc: 0.811465	training's binary_logloss: 0.0828583	valid_1's auc: 0.823632	valid_1's binary_logloss: 0.0599532
[109]	training's auc: 0.811584	training's binary_logloss: 0.0828442	valid_1's auc: 0.823672	valid_1's binary_logloss: 0.0599447
[110]	training's auc: 0.811635	training's binary_logloss: 0.0828381	valid_1's auc: 0.82368	valid_1's binary_logloss: 0.059941
[111]	training's auc: 0.811691	training's binary_logloss: 0.0828319	valid_1's auc: 0.823625	valid_1's binary_logloss: 0.0599361
[112]	training's auc: 0.811854	training's binary_logloss: 0.0828229	valid_1's auc: 0.823589	valid_1's binary_logloss: 0.0599318
[113]	training's auc: 0.811929	training's binary_logloss: 0.0828131	valid_1's auc: 0.823237	valid_1's binary_logloss: 0.0599332
[114]	training's auc: 0.81202	training's binary_logloss: 0.0828053	valid_1's auc: 0.822993	valid_1's binary_logloss: 0.0599489
[115]	training's auc: 0.812045	training's binary_logloss: 0.0827987	valid_1's auc: 0.823006	valid_1's binary_logloss: 0.0599442
[116]	training's auc: 0.812097	training's binary_logloss: 0.0827935	valid_1's auc: 0.823076	valid_1's binary_logloss: 0.0599337
[117]	training's auc: 0.812219	training's binary_logloss: 0.0827841	valid_1's auc: 0.823066	valid_1's binary_logloss: 0.0599397
[118]	training's auc: 0.812252	training's binary_logloss: 0.0827784	valid_1's auc: 0.82313	valid_1's binary_logloss: 0.0599415
[119]	training's auc: 0.812329	training's binary_logloss: 0.0827701	valid_1's auc: 0.823238	valid_1's binary_logloss: 0.0599377
[120]	training's auc: 0.812437	training's binary_logloss: 0.0827596	valid_1's auc: 0.823315	valid_1's binary_logloss: 0.0599267
[121]	training's auc: 0.812574	training's binary_logloss: 0.0827483	valid_1's auc: 0.823122	valid_1's binary_logloss: 0.05994
[122]	training's auc: 0.812613	training's binary_logloss: 0.0827436	valid_1's auc: 0.822983	valid_1's binary_logloss: 0.0599483
[123]	training's auc: 0.812644	training's binary_logloss: 0.0827392	valid_1's auc: 0.822837	valid_1's binary_logloss: 0.0599535
[124]	training's auc: 0.812722	training's binary_logloss: 0.0827306	valid_1's auc: 0.82294	valid_1's binary_logloss: 0.059952
[125]	training's auc: 0.812771	training's binary_logloss: 0.0827202	valid_1's auc: 0.822776	valid_1's binary_logloss: 0.059963
[126]	training's auc: 0.812805	training's binary_logloss: 0.0827155	valid_1's auc: 0.822784	valid_1's binary_logloss: 0.0599671
[127]	training's auc: 0.812825	training's binary_logloss: 0.082712	valid_1's auc: 0.822783	valid_1's binary_logloss: 0.0599627
[128]	training's auc: 0.812853	training's binary_logloss: 0.0827075	valid_1's auc: 0.822657	valid_1's binary_logloss: 0.0599685
[129]	training's auc: 0.812864	training's binary_logloss: 0.0827033	valid_1's auc: 0.822662	valid_1's binary_logloss: 0.0599673
[130]	training's auc: 0.812904	training's binary_logloss: 0.0826986	valid_1's auc: 0.82258	valid_1's binary_logloss: 0.0599632
[131]	training's auc: 0.81296	training's binary_logloss: 0.0826931	valid_1's auc: 0.822559	valid_1's binary_logloss: 0.0599625
[132]	training's auc: 0.812977	training's binary_logloss: 0.0826885	valid_1's auc: 0.822703	valid_1's binary_logloss: 0.0599417
[133]	training's auc: 0.813021	training's binary_logloss: 0.0826841	valid_1's auc: 0.822615	valid_1's binary_logloss: 0.0599482
[134]	training's auc: 0.813034	training's binary_logloss: 0.0826782	valid_1's auc: 0.822601	valid_1's binary_logloss: 0.0599442
[135]	training's auc: 0.813064	training's binary_logloss: 0.082674	valid_1's auc: 0.822597	valid_1's binary_logloss: 0.0599422
[136]	training's auc: 0.813152	training's binary_logloss: 0.0826665	valid_1's auc: 0.82274	valid_1's binary_logloss: 0.0599408
[137]	training's auc: 0.813161	training's binary_logloss: 0.0826624	valid_1's auc: 0.822848	valid_1's binary_logloss: 0.0599344
[138]	training's auc: 0.813177	training's binary_logloss: 0.0826574	valid_1's auc: 0.822831	valid_1's binary_logloss: 0.059931
[139]	training's auc: 0.813186	training's binary_logloss: 0.0826547	valid_1's auc: 0.822759	valid_1's binary_logloss: 0.0599297
[140]	training's auc: 0.813299	training's binary_logloss: 0.0826421	valid_1's auc: 0.822658	valid_1's binary_logloss: 0.0599483
[141]	training's auc: 0.813336	training's binary_logloss: 0.0826374	valid_1's auc: 0.822697	valid_1's binary_logloss: 0.0599498
[142]	training's auc: 0.813438	training's binary_logloss: 0.0826291	valid_1's auc: 0.822632	valid_1's binary_logloss: 0.059949
[143]	training's auc: 0.813443	training's binary_logloss: 0.0826246	valid_1's auc: 0.822654	valid_1's binary_logloss: 0.0599424
[144]	training's auc: 0.813508	training's binary_logloss: 0.0826121	valid_1's auc: 0.822624	valid_1's binary_logloss: 0.0599503
[145]	training's auc: 0.813552	training's binary_logloss: 0.0826029	valid_1's auc: 0.82264	valid_1's binary_logloss: 0.059951
[146]	training's auc: 0.813679	training's binary_logloss: 0.082593	valid_1's auc: 0.822462	valid_1's binary_logloss: 0.0599591
[147]	training's auc: 0.813806	training's binary_logloss: 0.082583	valid_1's auc: 0.822468	valid_1's binary_logloss: 0.0599645
[148]	training's auc: 0.8139	training's binary_logloss: 0.0825734	valid_1's auc: 0.822298	valid_1's binary_logloss: 0.0599615
[149]	training's auc: 0.813913	training's binary_logloss: 0.0825699	valid_1's auc: 0.822275	valid_1's binary_logloss: 0.0599551
[150]	training's auc: 0.81395	training's binary_logloss: 0.0825639	valid_1's auc: 0.822281	valid_1's binary_logloss: 0.0599576
[151]	training's auc: 0.81404	training's binary_logloss: 0.0825578	valid_1's auc: 0.82229	valid_1's binary_logloss: 0.0599587
[152]	training's auc: 0.814013	training's binary_logloss: 0.0825556	valid_1's auc: 0.822147	valid_1's binary_logloss: 0.0599682
[153]	training's auc: 0.814033	training's binary_logloss: 0.0825419	valid_1's auc: 0.822199	valid_1's binary_logloss: 0.0599495
Early stopping, best iteration is:
[53]	training's auc: 0.806651	training's binary_logloss: 0.0838034	valid_1's auc: 0.825846	valid_1's binary_logloss: 0.060648
8
[1]	training's auc: 0.655534	training's binary_logloss: 0.0921053	valid_1's auc: 0.64957	valid_1's binary_logloss: 0.0890446
Training until validation scores don't improve for 100 rounds.
[2]	training's auc: 0.784031	training's binary_logloss: 0.0906779	valid_1's auc: 0.754556	valid_1's binary_logloss: 0.0879805
[3]	training's auc: 0.792225	training's binary_logloss: 0.0900435	valid_1's auc: 0.756691	valid_1's binary_logloss: 0.0874868
[4]	training's auc: 0.792269	training's binary_logloss: 0.0894428	valid_1's auc: 0.762152	valid_1's binary_logloss: 0.0869681
[5]	training's auc: 0.800292	training's binary_logloss: 0.0884055	valid_1's auc: 0.76791	valid_1's binary_logloss: 0.086218
[6]	training's auc: 0.806541	training's binary_logloss: 0.0874556	valid_1's auc: 0.773433	valid_1's binary_logloss: 0.0855461
[7]	training's auc: 0.80569	training's binary_logloss: 0.0870122	valid_1's auc: 0.77362	valid_1's binary_logloss: 0.0851024
[8]	training's auc: 0.805672	training's binary_logloss: 0.0862701	valid_1's auc: 0.775316	valid_1's binary_logloss: 0.0845559
[9]	training's auc: 0.808261	training's binary_logloss: 0.0860663	valid_1's auc: 0.779628	valid_1's binary_logloss: 0.0843609
[10]	training's auc: 0.807458	training's binary_logloss: 0.0857178	valid_1's auc: 0.781358	valid_1's binary_logloss: 0.0840543
[11]	training's auc: 0.808021	training's binary_logloss: 0.0850958	valid_1's auc: 0.781175	valid_1's binary_logloss: 0.0836284
[12]	training's auc: 0.810203	training's binary_logloss: 0.0845284	valid_1's auc: 0.782227	valid_1's binary_logloss: 0.0832315
[13]	training's auc: 0.810169	training's binary_logloss: 0.0842948	valid_1's auc: 0.780961	valid_1's binary_logloss: 0.0830693
[14]	training's auc: 0.809282	training's binary_logloss: 0.0840578	valid_1's auc: 0.781346	valid_1's binary_logloss: 0.0828613
[15]	training's auc: 0.810176	training's binary_logloss: 0.0836058	valid_1's auc: 0.781528	valid_1's binary_logloss: 0.0825443
[16]	training's auc: 0.811237	training's binary_logloss: 0.0832011	valid_1's auc: 0.782219	valid_1's binary_logloss: 0.0822844
[17]	training's auc: 0.810978	training's binary_logloss: 0.0829828	valid_1's auc: 0.782345	valid_1's binary_logloss: 0.0820686
[18]	training's auc: 0.811253	training's binary_logloss: 0.0826373	valid_1's auc: 0.782161	valid_1's binary_logloss: 0.0818394
[19]	training's auc: 0.812154	training's binary_logloss: 0.0823514	valid_1's auc: 0.782977	valid_1's binary_logloss: 0.081622
[20]	training's auc: 0.811921	training's binary_logloss: 0.0821776	valid_1's auc: 0.782479	valid_1's binary_logloss: 0.0814714
[21]	training's auc: 0.812247	training's binary_logloss: 0.081933	valid_1's auc: 0.782736	valid_1's binary_logloss: 0.0813235
[22]	training's auc: 0.812982	training's binary_logloss: 0.0816813	valid_1's auc: 0.782961	valid_1's binary_logloss: 0.0811561
[23]	training's auc: 0.813761	training's binary_logloss: 0.0815328	valid_1's auc: 0.782592	valid_1's binary_logloss: 0.0810227
[24]	training's auc: 0.813515	training's binary_logloss: 0.081318	valid_1's auc: 0.78257	valid_1's binary_logloss: 0.0808831
[25]	training's auc: 0.813456	training's binary_logloss: 0.0811932	valid_1's auc: 0.782132	valid_1's binary_logloss: 0.080776
[26]	training's auc: 0.814143	training's binary_logloss: 0.0810232	valid_1's auc: 0.783143	valid_1's binary_logloss: 0.080685
[27]	training's auc: 0.814392	training's binary_logloss: 0.0808586	valid_1's auc: 0.782958	valid_1's binary_logloss: 0.0805982
[28]	training's auc: 0.815114	training's binary_logloss: 0.0807799	valid_1's auc: 0.783904	valid_1's binary_logloss: 0.0805204
[29]	training's auc: 0.815326	training's binary_logloss: 0.0806284	valid_1's auc: 0.783214	valid_1's binary_logloss: 0.0804473
[30]	training's auc: 0.815445	training's binary_logloss: 0.0805127	valid_1's auc: 0.783504	valid_1's binary_logloss: 0.080337
[31]	training's auc: 0.815664	training's binary_logloss: 0.0803734	valid_1's auc: 0.783092	valid_1's binary_logloss: 0.080245
[32]	training's auc: 0.815461	training's binary_logloss: 0.0803042	valid_1's auc: 0.783076	valid_1's binary_logloss: 0.0801969
[33]	training's auc: 0.815944	training's binary_logloss: 0.0801849	valid_1's auc: 0.78335	valid_1's binary_logloss: 0.0801189
[34]	training's auc: 0.816011	training's binary_logloss: 0.0801248	valid_1's auc: 0.783221	valid_1's binary_logloss: 0.0800747
[35]	training's auc: 0.816264	training's binary_logloss: 0.0800343	valid_1's auc: 0.78347	valid_1's binary_logloss: 0.0800399
[36]	training's auc: 0.816413	training's binary_logloss: 0.0799276	valid_1's auc: 0.783579	valid_1's binary_logloss: 0.0799957
[37]	training's auc: 0.816717	training's binary_logloss: 0.0798412	valid_1's auc: 0.783838	valid_1's binary_logloss: 0.0799309
[38]	training's auc: 0.816858	training's binary_logloss: 0.0797467	valid_1's auc: 0.783795	valid_1's binary_logloss: 0.0798736
[39]	training's auc: 0.817364	training's binary_logloss: 0.0796987	valid_1's auc: 0.783824	valid_1's binary_logloss: 0.0798193
[40]	training's auc: 0.817382	training's binary_logloss: 0.0796192	valid_1's auc: 0.783799	valid_1's binary_logloss: 0.0797769
[41]	training's auc: 0.817685	training's binary_logloss: 0.0795346	valid_1's auc: 0.784231	valid_1's binary_logloss: 0.0797304
[42]	training's auc: 0.817907	training's binary_logloss: 0.0794741	valid_1's auc: 0.78407	valid_1's binary_logloss: 0.0797124
[43]	training's auc: 0.818201	training's binary_logloss: 0.0794205	valid_1's auc: 0.784022	valid_1's binary_logloss: 0.0796767
[44]	training's auc: 0.818218	training's binary_logloss: 0.0793631	valid_1's auc: 0.784115	valid_1's binary_logloss: 0.079661
[45]	training's auc: 0.818524	training's binary_logloss: 0.0792982	valid_1's auc: 0.783571	valid_1's binary_logloss: 0.0796149
[46]	training's auc: 0.818555	training's binary_logloss: 0.0792513	valid_1's auc: 0.783648	valid_1's binary_logloss: 0.0795808
[47]	training's auc: 0.818615	training's binary_logloss: 0.0791979	valid_1's auc: 0.783478	valid_1's binary_logloss: 0.0795826
[48]	training's auc: 0.818592	training's binary_logloss: 0.0791522	valid_1's auc: 0.783445	valid_1's binary_logloss: 0.0795619
[49]	training's auc: 0.818876	training's binary_logloss: 0.0791164	valid_1's auc: 0.78382	valid_1's binary_logloss: 0.0795243
[50]	training's auc: 0.819024	training's binary_logloss: 0.0790847	valid_1's auc: 0.783502	valid_1's binary_logloss: 0.0795269
[51]	training's auc: 0.818982	training's binary_logloss: 0.0790376	valid_1's auc: 0.783484	valid_1's binary_logloss: 0.0794939
[52]	training's auc: 0.818955	training's binary_logloss: 0.0790008	valid_1's auc: 0.783276	valid_1's binary_logloss: 0.0794859
[53]	training's auc: 0.819034	training's binary_logloss: 0.078982	valid_1's auc: 0.783205	valid_1's binary_logloss: 0.0794826
[54]	training's auc: 0.819007	training's binary_logloss: 0.0789487	valid_1's auc: 0.783178	valid_1's binary_logloss: 0.0794655
[55]	training's auc: 0.819106	training's binary_logloss: 0.0789174	valid_1's auc: 0.783313	valid_1's binary_logloss: 0.0794325
[56]	training's auc: 0.818988	training's binary_logloss: 0.0788839	valid_1's auc: 0.783629	valid_1's binary_logloss: 0.0794011
[57]	training's auc: 0.819188	training's binary_logloss: 0.0788547	valid_1's auc: 0.783817	valid_1's binary_logloss: 0.0793864
[58]	training's auc: 0.819402	training's binary_logloss: 0.0788221	valid_1's auc: 0.783963	valid_1's binary_logloss: 0.0793541
[59]	training's auc: 0.81949	training's binary_logloss: 0.0787999	valid_1's auc: 0.784072	valid_1's binary_logloss: 0.079331
[60]	training's auc: 0.819589	training's binary_logloss: 0.0787798	valid_1's auc: 0.784306	valid_1's binary_logloss: 0.0793165
[61]	training's auc: 0.819728	training's binary_logloss: 0.0787515	valid_1's auc: 0.784216	valid_1's binary_logloss: 0.0792937
[62]	training's auc: 0.819662	training's binary_logloss: 0.0787233	valid_1's auc: 0.784164	valid_1's binary_logloss: 0.0792801
[63]	training's auc: 0.819813	training's binary_logloss: 0.0787029	valid_1's auc: 0.784571	valid_1's binary_logloss: 0.0792499
[64]	training's auc: 0.819947	training's binary_logloss: 0.0786747	valid_1's auc: 0.78446	valid_1's binary_logloss: 0.0792533
[65]	training's auc: 0.820022	training's binary_logloss: 0.0786538	valid_1's auc: 0.784472	valid_1's binary_logloss: 0.0792316
[66]	training's auc: 0.820127	training's binary_logloss: 0.0786348	valid_1's auc: 0.784664	valid_1's binary_logloss: 0.0792305
[67]	training's auc: 0.820297	training's binary_logloss: 0.0786163	valid_1's auc: 0.784947	valid_1's binary_logloss: 0.0791983
[68]	training's auc: 0.820432	training's binary_logloss: 0.0785906	valid_1's auc: 0.785215	valid_1's binary_logloss: 0.0791634
[69]	training's auc: 0.820567	training's binary_logloss: 0.0785762	valid_1's auc: 0.785291	valid_1's binary_logloss: 0.0791573
[70]	training's auc: 0.820656	training's binary_logloss: 0.078559	valid_1's auc: 0.785022	valid_1's binary_logloss: 0.0791504
[71]	training's auc: 0.820772	training's binary_logloss: 0.0785481	valid_1's auc: 0.785291	valid_1's binary_logloss: 0.079121
[72]	training's auc: 0.820776	training's binary_logloss: 0.0785264	valid_1's auc: 0.785388	valid_1's binary_logloss: 0.0791141
[73]	training's auc: 0.820885	training's binary_logloss: 0.0785048	valid_1's auc: 0.785601	valid_1's binary_logloss: 0.0790928
[74]	training's auc: 0.820993	training's binary_logloss: 0.0784918	valid_1's auc: 0.785674	valid_1's binary_logloss: 0.0790998
[75]	training's auc: 0.821025	training's binary_logloss: 0.0784787	valid_1's auc: 0.785609	valid_1's binary_logloss: 0.0791091
[76]	training's auc: 0.821143	training's binary_logloss: 0.0784582	valid_1's auc: 0.785533	valid_1's binary_logloss: 0.0790983
[77]	training's auc: 0.821153	training's binary_logloss: 0.0784449	valid_1's auc: 0.785453	valid_1's binary_logloss: 0.0790925
[78]	training's auc: 0.821267	training's binary_logloss: 0.0784375	valid_1's auc: 0.785365	valid_1's binary_logloss: 0.0790976
[79]	training's auc: 0.821337	training's binary_logloss: 0.0784289	valid_1's auc: 0.785413	valid_1's binary_logloss: 0.0791004
[80]	training's auc: 0.821432	training's binary_logloss: 0.0784187	valid_1's auc: 0.785352	valid_1's binary_logloss: 0.0791017
[81]	training's auc: 0.821432	training's binary_logloss: 0.0784016	valid_1's auc: 0.78561	valid_1's binary_logloss: 0.0790911
[82]	training's auc: 0.821599	training's binary_logloss: 0.0783875	valid_1's auc: 0.785593	valid_1's binary_logloss: 0.0790856
[83]	training's auc: 0.821741	training's binary_logloss: 0.0783701	valid_1's auc: 0.785815	valid_1's binary_logloss: 0.0790796
[84]	training's auc: 0.821787	training's binary_logloss: 0.0783598	valid_1's auc: 0.785932	valid_1's binary_logloss: 0.0790788
[85]	training's auc: 0.821866	training's binary_logloss: 0.0783494	valid_1's auc: 0.786283	valid_1's binary_logloss: 0.0790641
[86]	training's auc: 0.821849	training's binary_logloss: 0.0783339	valid_1's auc: 0.786232	valid_1's binary_logloss: 0.079058
[87]	training's auc: 0.821839	training's binary_logloss: 0.0783175	valid_1's auc: 0.785939	valid_1's binary_logloss: 0.0790535
[88]	training's auc: 0.821931	training's binary_logloss: 0.0783039	valid_1's auc: 0.78613	valid_1's binary_logloss: 0.0790461
[89]	training's auc: 0.821914	training's binary_logloss: 0.0782978	valid_1's auc: 0.786344	valid_1's binary_logloss: 0.0790378
[90]	training's auc: 0.821963	training's binary_logloss: 0.0782853	valid_1's auc: 0.786261	valid_1's binary_logloss: 0.0790495
[91]	training's auc: 0.822054	training's binary_logloss: 0.0782734	valid_1's auc: 0.786616	valid_1's binary_logloss: 0.0790317
[92]	training's auc: 0.822196	training's binary_logloss: 0.078264	valid_1's auc: 0.78647	valid_1's binary_logloss: 0.0790418
[93]	training's auc: 0.82227	training's binary_logloss: 0.0782511	valid_1's auc: 0.786457	valid_1's binary_logloss: 0.0790237
[94]	training's auc: 0.822332	training's binary_logloss: 0.0782432	valid_1's auc: 0.78645	valid_1's binary_logloss: 0.0790311
[95]	training's auc: 0.822368	training's binary_logloss: 0.0782334	valid_1's auc: 0.786355	valid_1's binary_logloss: 0.0790374
[96]	training's auc: 0.822411	training's binary_logloss: 0.0782272	valid_1's auc: 0.786398	valid_1's binary_logloss: 0.0790194
[97]	training's auc: 0.822502	training's binary_logloss: 0.0782184	valid_1's auc: 0.78646	valid_1's binary_logloss: 0.079007
[98]	training's auc: 0.822563	training's binary_logloss: 0.0782044	valid_1's auc: 0.786401	valid_1's binary_logloss: 0.0789917
[99]	training's auc: 0.822627	training's binary_logloss: 0.0781944	valid_1's auc: 0.786408	valid_1's binary_logloss: 0.0789963
[100]	training's auc: 0.822656	training's binary_logloss: 0.078188	valid_1's auc: 0.786307	valid_1's binary_logloss: 0.0790026
[101]	training's auc: 0.822742	training's binary_logloss: 0.0781771	valid_1's auc: 0.786278	valid_1's binary_logloss: 0.0790049
[102]	training's auc: 0.82282	training's binary_logloss: 0.0781672	valid_1's auc: 0.786192	valid_1's binary_logloss: 0.0790083
[103]	training's auc: 0.822795	training's binary_logloss: 0.0781576	valid_1's auc: 0.786247	valid_1's binary_logloss: 0.0790055
[104]	training's auc: 0.822816	training's binary_logloss: 0.0781485	valid_1's auc: 0.786197	valid_1's binary_logloss: 0.07901
[105]	training's auc: 0.822842	training's binary_logloss: 0.0781402	valid_1's auc: 0.786148	valid_1's binary_logloss: 0.0790222
[106]	training's auc: 0.822924	training's binary_logloss: 0.0781312	valid_1's auc: 0.786128	valid_1's binary_logloss: 0.0790196
[107]	training's auc: 0.822985	training's binary_logloss: 0.0781237	valid_1's auc: 0.786337	valid_1's binary_logloss: 0.07901
[108]	training's auc: 0.823039	training's binary_logloss: 0.0781189	valid_1's auc: 0.786351	valid_1's binary_logloss: 0.0790103
[109]	training's auc: 0.823078	training's binary_logloss: 0.0781112	valid_1's auc: 0.786267	valid_1's binary_logloss: 0.0790051
[110]	training's auc: 0.82319	training's binary_logloss: 0.0781041	valid_1's auc: 0.786363	valid_1's binary_logloss: 0.0790121
[111]	training's auc: 0.823236	training's binary_logloss: 0.0780987	valid_1's auc: 0.786377	valid_1's binary_logloss: 0.0790151
[112]	training's auc: 0.823316	training's binary_logloss: 0.0780912	valid_1's auc: 0.786322	valid_1's binary_logloss: 0.0790217
[113]	training's auc: 0.823261	training's binary_logloss: 0.0780737	valid_1's auc: 0.786354	valid_1's binary_logloss: 0.0790061
[114]	training's auc: 0.823326	training's binary_logloss: 0.0780691	valid_1's auc: 0.786429	valid_1's binary_logloss: 0.0790017
[115]	training's auc: 0.823345	training's binary_logloss: 0.0780657	valid_1's auc: 0.78652	valid_1's binary_logloss: 0.078992
[116]	training's auc: 0.823398	training's binary_logloss: 0.0780629	valid_1's auc: 0.786453	valid_1's binary_logloss: 0.078996
[117]	training's auc: 0.82352	training's binary_logloss: 0.078055	valid_1's auc: 0.786324	valid_1's binary_logloss: 0.0789912
[118]	training's auc: 0.823639	training's binary_logloss: 0.0780459	valid_1's auc: 0.786357	valid_1's binary_logloss: 0.0789927
[119]	training's auc: 0.823765	training's binary_logloss: 0.0780379	valid_1's auc: 0.786363	valid_1's binary_logloss: 0.078994
[120]	training's auc: 0.823795	training's binary_logloss: 0.0780322	valid_1's auc: 0.786589	valid_1's binary_logloss: 0.0789764
[121]	training's auc: 0.82394	training's binary_logloss: 0.0780226	valid_1's auc: 0.786529	valid_1's binary_logloss: 0.0789795
[122]	training's auc: 0.823982	training's binary_logloss: 0.0780177	valid_1's auc: 0.786606	valid_1's binary_logloss: 0.0789843
[123]	training's auc: 0.82402	training's binary_logloss: 0.0780125	valid_1's auc: 0.786626	valid_1's binary_logloss: 0.0789717
[124]	training's auc: 0.824026	training's binary_logloss: 0.0780057	valid_1's auc: 0.786676	valid_1's binary_logloss: 0.0789752
[125]	training's auc: 0.824058	training's binary_logloss: 0.0780015	valid_1's auc: 0.786836	valid_1's binary_logloss: 0.0789708
[126]	training's auc: 0.824083	training's binary_logloss: 0.0779982	valid_1's auc: 0.786917	valid_1's binary_logloss: 0.0789604
[127]	training's auc: 0.824091	training's binary_logloss: 0.0779944	valid_1's auc: 0.787011	valid_1's binary_logloss: 0.0789621
[128]	training's auc: 0.824093	training's binary_logloss: 0.0779903	valid_1's auc: 0.786994	valid_1's binary_logloss: 0.0789605
[129]	training's auc: 0.824115	training's binary_logloss: 0.0779862	valid_1's auc: 0.786917	valid_1's binary_logloss: 0.0789549
[130]	training's auc: 0.824134	training's binary_logloss: 0.07798	valid_1's auc: 0.786869	valid_1's binary_logloss: 0.0789555
[131]	training's auc: 0.824169	training's binary_logloss: 0.0779732	valid_1's auc: 0.786889	valid_1's binary_logloss: 0.0789679
[132]	training's auc: 0.824176	training's binary_logloss: 0.0779674	valid_1's auc: 0.786952	valid_1's binary_logloss: 0.0789678
[133]	training's auc: 0.824187	training's binary_logloss: 0.077964	valid_1's auc: 0.787101	valid_1's binary_logloss: 0.0789585
[134]	training's auc: 0.824254	training's binary_logloss: 0.0779584	valid_1's auc: 0.787027	valid_1's binary_logloss: 0.0789541
[135]	training's auc: 0.824341	training's binary_logloss: 0.077952	valid_1's auc: 0.786869	valid_1's binary_logloss: 0.0789723
[136]	training's auc: 0.82438	training's binary_logloss: 0.0779474	valid_1's auc: 0.78692	valid_1's binary_logloss: 0.0789597
[137]	training's auc: 0.824406	training's binary_logloss: 0.0779433	valid_1's auc: 0.786874	valid_1's binary_logloss: 0.0789575
[138]	training's auc: 0.824454	training's binary_logloss: 0.0779352	valid_1's auc: 0.786895	valid_1's binary_logloss: 0.0789616
[139]	training's auc: 0.824473	training's binary_logloss: 0.0779328	valid_1's auc: 0.786893	valid_1's binary_logloss: 0.0789637
[140]	training's auc: 0.82449	training's binary_logloss: 0.0779292	valid_1's auc: 0.786926	valid_1's binary_logloss: 0.0789555
[141]	training's auc: 0.824514	training's binary_logloss: 0.0779257	valid_1's auc: 0.786878	valid_1's binary_logloss: 0.0789611
[142]	training's auc: 0.824559	training's binary_logloss: 0.0779182	valid_1's auc: 0.786888	valid_1's binary_logloss: 0.0789665
[143]	training's auc: 0.824643	training's binary_logloss: 0.0779107	valid_1's auc: 0.786796	valid_1's binary_logloss: 0.0789839
[144]	training's auc: 0.824719	training's binary_logloss: 0.0779031	valid_1's auc: 0.786829	valid_1's binary_logloss: 0.078982
[145]	training's auc: 0.824727	training's binary_logloss: 0.0778948	valid_1's auc: 0.786816	valid_1's binary_logloss: 0.078985
[146]	training's auc: 0.824755	training's binary_logloss: 0.0778917	valid_1's auc: 0.786763	valid_1's binary_logloss: 0.0789806
[147]	training's auc: 0.824772	training's binary_logloss: 0.0778878	valid_1's auc: 0.786758	valid_1's binary_logloss: 0.078983
[148]	training's auc: 0.824852	training's binary_logloss: 0.0778802	valid_1's auc: 0.786854	valid_1's binary_logloss: 0.0789847
[149]	training's auc: 0.824865	training's binary_logloss: 0.077876	valid_1's auc: 0.786902	valid_1's binary_logloss: 0.0789814
[150]	training's auc: 0.824915	training's binary_logloss: 0.0778703	valid_1's auc: 0.786859	valid_1's binary_logloss: 0.0789805
[151]	training's auc: 0.824947	training's binary_logloss: 0.0778578	valid_1's auc: 0.786807	valid_1's binary_logloss: 0.0789573
[152]	training's auc: 0.824974	training's binary_logloss: 0.0778552	valid_1's auc: 0.786779	valid_1's binary_logloss: 0.0789618
[153]	training's auc: 0.824993	training's binary_logloss: 0.0778495	valid_1's auc: 0.786761	valid_1's binary_logloss: 0.0789768
[154]	training's auc: 0.825031	training's binary_logloss: 0.0778448	valid_1's auc: 0.786767	valid_1's binary_logloss: 0.0789732
[155]	training's auc: 0.825142	training's binary_logloss: 0.0778363	valid_1's auc: 0.786753	valid_1's binary_logloss: 0.0789691
[156]	training's auc: 0.825185	training's binary_logloss: 0.0778336	valid_1's auc: 0.786824	valid_1's binary_logloss: 0.0789662
[157]	training's auc: 0.825244	training's binary_logloss: 0.077829	valid_1's auc: 0.786973	valid_1's binary_logloss: 0.078953
[158]	training's auc: 0.825257	training's binary_logloss: 0.0778269	valid_1's auc: 0.786975	valid_1's binary_logloss: 0.0789466
[159]	training's auc: 0.825286	training's binary_logloss: 0.077823	valid_1's auc: 0.78702	valid_1's binary_logloss: 0.0789322
[160]	training's auc: 0.825357	training's binary_logloss: 0.0778159	valid_1's auc: 0.78703	valid_1's binary_logloss: 0.0789328
[161]	training's auc: 0.825411	training's binary_logloss: 0.0778119	valid_1's auc: 0.78692	valid_1's binary_logloss: 0.0789383
[162]	training's auc: 0.825434	training's binary_logloss: 0.0778067	valid_1's auc: 0.786873	valid_1's binary_logloss: 0.0789497
[163]	training's auc: 0.825453	training's binary_logloss: 0.0778046	valid_1's auc: 0.786864	valid_1's binary_logloss: 0.0789537
[164]	training's auc: 0.825503	training's binary_logloss: 0.077798	valid_1's auc: 0.786776	valid_1's binary_logloss: 0.0789576
[165]	training's auc: 0.825541	training's binary_logloss: 0.0777844	valid_1's auc: 0.786773	valid_1's binary_logloss: 0.0789516
[166]	training's auc: 0.825539	training's binary_logloss: 0.0777775	valid_1's auc: 0.786653	valid_1's binary_logloss: 0.078943
[167]	training's auc: 0.825606	training's binary_logloss: 0.0777724	valid_1's auc: 0.786586	valid_1's binary_logloss: 0.0789509
[168]	training's auc: 0.825606	training's binary_logloss: 0.077771	valid_1's auc: 0.786555	valid_1's binary_logloss: 0.0789529
[169]	training's auc: 0.825623	training's binary_logloss: 0.0777661	valid_1's auc: 0.786428	valid_1's binary_logloss: 0.0789606
[170]	training's auc: 0.825644	training's binary_logloss: 0.0777637	valid_1's auc: 0.786396	valid_1's binary_logloss: 0.0789655
[171]	training's auc: 0.825668	training's binary_logloss: 0.0777602	valid_1's auc: 0.786352	valid_1's binary_logloss: 0.0789625
[172]	training's auc: 0.825707	training's binary_logloss: 0.0777449	valid_1's auc: 0.786399	valid_1's binary_logloss: 0.07897
[173]	training's auc: 0.825756	training's binary_logloss: 0.077737	valid_1's auc: 0.78632	valid_1's binary_logloss: 0.0789739
[174]	training's auc: 0.825777	training's binary_logloss: 0.077735	valid_1's auc: 0.786218	valid_1's binary_logloss: 0.0789728
[175]	training's auc: 0.825797	training's binary_logloss: 0.0777302	valid_1's auc: 0.786371	valid_1's binary_logloss: 0.0789586
[176]	training's auc: 0.82586	training's binary_logloss: 0.0777229	valid_1's auc: 0.786413	valid_1's binary_logloss: 0.0789549
[177]	training's auc: 0.82586	training's binary_logloss: 0.0777215	valid_1's auc: 0.786354	valid_1's binary_logloss: 0.0789554
[178]	training's auc: 0.825978	training's binary_logloss: 0.0777139	valid_1's auc: 0.786282	valid_1's binary_logloss: 0.0789642
[179]	training's auc: 0.826004	training's binary_logloss: 0.077712	valid_1's auc: 0.78624	valid_1's binary_logloss: 0.0789668
[180]	training's auc: 0.826015	training's binary_logloss: 0.0777086	valid_1's auc: 0.786241	valid_1's binary_logloss: 0.0789582
[181]	training's auc: 0.826025	training's binary_logloss: 0.0777058	valid_1's auc: 0.786223	valid_1's binary_logloss: 0.0789672
[182]	training's auc: 0.826054	training's binary_logloss: 0.0777037	valid_1's auc: 0.786208	valid_1's binary_logloss: 0.0789682
[183]	training's auc: 0.826064	training's binary_logloss: 0.0777011	valid_1's auc: 0.786239	valid_1's binary_logloss: 0.0789613
[184]	training's auc: 0.826124	training's binary_logloss: 0.0776982	valid_1's auc: 0.786293	valid_1's binary_logloss: 0.0789611
[185]	training's auc: 0.826139	training's binary_logloss: 0.0776903	valid_1's auc: 0.786263	valid_1's binary_logloss: 0.0789511
[186]	training's auc: 0.826179	training's binary_logloss: 0.0776848	valid_1's auc: 0.786334	valid_1's binary_logloss: 0.078947
[187]	training's auc: 0.826215	training's binary_logloss: 0.0776789	valid_1's auc: 0.786401	valid_1's binary_logloss: 0.0789423
[188]	training's auc: 0.826235	training's binary_logloss: 0.077676	valid_1's auc: 0.786488	valid_1's binary_logloss: 0.0789343
[189]	training's auc: 0.826282	training's binary_logloss: 0.0776715	valid_1's auc: 0.786471	valid_1's binary_logloss: 0.0789407
[190]	training's auc: 0.826278	training's binary_logloss: 0.0776701	valid_1's auc: 0.786426	valid_1's binary_logloss: 0.0789375
[191]	training's auc: 0.826325	training's binary_logloss: 0.0776637	valid_1's auc: 0.786376	valid_1's binary_logloss: 0.0789415
[192]	training's auc: 0.826328	training's binary_logloss: 0.0776618	valid_1's auc: 0.786346	valid_1's binary_logloss: 0.0789386
[193]	training's auc: 0.826437	training's binary_logloss: 0.0776509	valid_1's auc: 0.786426	valid_1's binary_logloss: 0.0789442
[194]	training's auc: 0.826493	training's binary_logloss: 0.0776317	valid_1's auc: 0.786441	valid_1's binary_logloss: 0.0789423
[195]	training's auc: 0.826518	training's binary_logloss: 0.0776291	valid_1's auc: 0.786308	valid_1's binary_logloss: 0.0789367
[196]	training's auc: 0.826582	training's binary_logloss: 0.0776236	valid_1's auc: 0.786229	valid_1's binary_logloss: 0.0789333
[197]	training's auc: 0.826606	training's binary_logloss: 0.0776201	valid_1's auc: 0.786179	valid_1's binary_logloss: 0.0789414
[198]	training's auc: 0.826633	training's binary_logloss: 0.0776174	valid_1's auc: 0.786172	valid_1's binary_logloss: 0.0789445
[199]	training's auc: 0.826654	training's binary_logloss: 0.0776125	valid_1's auc: 0.786157	valid_1's binary_logloss: 0.0789483
[200]	training's auc: 0.826726	training's binary_logloss: 0.0776032	valid_1's auc: 0.786308	valid_1's binary_logloss: 0.0789347
[201]	training's auc: 0.826736	training's binary_logloss: 0.0776011	valid_1's auc: 0.786306	valid_1's binary_logloss: 0.07893
[202]	training's auc: 0.826725	training's binary_logloss: 0.0775985	valid_1's auc: 0.78634	valid_1's binary_logloss: 0.0789325
[203]	training's auc: 0.826761	training's binary_logloss: 0.0775836	valid_1's auc: 0.786332	valid_1's binary_logloss: 0.0789319
[204]	training's auc: 0.826767	training's binary_logloss: 0.0775756	valid_1's auc: 0.786323	valid_1's binary_logloss: 0.0789358
[205]	training's auc: 0.826826	training's binary_logloss: 0.0775704	valid_1's auc: 0.78636	valid_1's binary_logloss: 0.0789362
[206]	training's auc: 0.826843	training's binary_logloss: 0.0775686	valid_1's auc: 0.786327	valid_1's binary_logloss: 0.0789443
[207]	training's auc: 0.826886	training's binary_logloss: 0.0775625	valid_1's auc: 0.786343	valid_1's binary_logloss: 0.0789469
[208]	training's auc: 0.826877	training's binary_logloss: 0.0775602	valid_1's auc: 0.786374	valid_1's binary_logloss: 0.0789392
[209]	training's auc: 0.826923	training's binary_logloss: 0.0775509	valid_1's auc: 0.786372	valid_1's binary_logloss: 0.0789281
[210]	training's auc: 0.826956	training's binary_logloss: 0.0775488	valid_1's auc: 0.786434	valid_1's binary_logloss: 0.0789274
[211]	training's auc: 0.827002	training's binary_logloss: 0.077542	valid_1's auc: 0.786457	valid_1's binary_logloss: 0.0789298
[212]	training's auc: 0.827065	training's binary_logloss: 0.0775354	valid_1's auc: 0.786537	valid_1's binary_logloss: 0.0789285
[213]	training's auc: 0.827099	training's binary_logloss: 0.077534	valid_1's auc: 0.78651	valid_1's binary_logloss: 0.0789412
[214]	training's auc: 0.827106	training's binary_logloss: 0.0775318	valid_1's auc: 0.786489	valid_1's binary_logloss: 0.078938
[215]	training's auc: 0.827147	training's binary_logloss: 0.0775288	valid_1's auc: 0.786432	valid_1's binary_logloss: 0.0789351
[216]	training's auc: 0.827218	training's binary_logloss: 0.0775226	valid_1's auc: 0.78638	valid_1's binary_logloss: 0.0789424
[217]	training's auc: 0.827248	training's binary_logloss: 0.0775197	valid_1's auc: 0.786353	valid_1's binary_logloss: 0.0789355
[218]	training's auc: 0.827266	training's binary_logloss: 0.0775178	valid_1's auc: 0.786303	valid_1's binary_logloss: 0.078938
[219]	training's auc: 0.827273	training's binary_logloss: 0.0775152	valid_1's auc: 0.786388	valid_1's binary_logloss: 0.0789348
[220]	training's auc: 0.827309	training's binary_logloss: 0.0775012	valid_1's auc: 0.786397	valid_1's binary_logloss: 0.0789287
[221]	training's auc: 0.827355	training's binary_logloss: 0.0774964	valid_1's auc: 0.786415	valid_1's binary_logloss: 0.0789237
[222]	training's auc: 0.827414	training's binary_logloss: 0.0774894	valid_1's auc: 0.786362	valid_1's binary_logloss: 0.0789336
[223]	training's auc: 0.827425	training's binary_logloss: 0.0774865	valid_1's auc: 0.786212	valid_1's binary_logloss: 0.0789445
[224]	training's auc: 0.827432	training's binary_logloss: 0.0774849	valid_1's auc: 0.786205	valid_1's binary_logloss: 0.0789539
[225]	training's auc: 0.827441	training's binary_logloss: 0.0774832	valid_1's auc: 0.786219	valid_1's binary_logloss: 0.0789549
[226]	training's auc: 0.827478	training's binary_logloss: 0.0774814	valid_1's auc: 0.786104	valid_1's binary_logloss: 0.0789654
[227]	training's auc: 0.827504	training's binary_logloss: 0.0774766	valid_1's auc: 0.786015	valid_1's binary_logloss: 0.078965
[228]	training's auc: 0.827541	training's binary_logloss: 0.0774724	valid_1's auc: 0.786071	valid_1's binary_logloss: 0.0789723
[229]	training's auc: 0.827635	training's binary_logloss: 0.0774622	valid_1's auc: 0.78603	valid_1's binary_logloss: 0.0789772
[230]	training's auc: 0.827692	training's binary_logloss: 0.0774549	valid_1's auc: 0.78611	valid_1's binary_logloss: 0.0789837
[231]	training's auc: 0.827744	training's binary_logloss: 0.0774514	valid_1's auc: 0.786075	valid_1's binary_logloss: 0.0789911
[232]	training's auc: 0.82773	training's binary_logloss: 0.0774475	valid_1's auc: 0.786088	valid_1's binary_logloss: 0.0789787
[233]	training's auc: 0.827737	training's binary_logloss: 0.0774452	valid_1's auc: 0.786147	valid_1's binary_logloss: 0.0789679
Early stopping, best iteration is:
[133]	training's auc: 0.824187	training's binary_logloss: 0.077964	valid_1's auc: 0.787101	valid_1's binary_logloss: 0.0789585
8
train_ks:  0.48924745315875107
test_ks:  0.47139484384477426
ft_lst:  ['finance_info', 'act_info', 'person_info', 'credit_info']
lst = ['person_info','finance_info','credit_info','act_info']

train = data[data.obs_mth != '2018-11-30'].reset_index().copy()
evl = data[data.obs_mth == '2018-11-30'].reset_index().copy()

x = train[lst]
y = train['bad_ind']

evl_x =  evl[lst]
evl_y = evl['bad_ind']

model,auc = LGB_test(x,y,evl_x,evl_y)

y_pred = model.predict_proba(x)[:,1]
fpr_lgb_train,tpr_lgb_train,_ = roc_curve(y,y_pred)
train_ks = abs(fpr_lgb_train - tpr_lgb_train).max()
print('train_ks : ',train_ks)

y_pred = model.predict_proba(evl_x)[:,1]
fpr_lgb,tpr_lgb,_ = roc_curve(evl_y,y_pred)
evl_ks = abs(fpr_lgb - tpr_lgb).max()
print('evl_ks : ',evl_ks)

from matplotlib import pyplot as plt
plt.plot(fpr_lgb_train,tpr_lgb_train,label = 'train LR')
plt.plot(fpr_lgb,tpr_lgb,label = 'evl LR')
plt.plot([0,1],[0,1],'k--')
plt.xlabel('False positive rate')
plt.ylabel('True positive rate')
plt.title('ROC Curve')
plt.legend(loc = 'best')
plt.show()
[1]	training's auc: 0.712883	training's binary_logloss: 0.0902863	valid_1's auc: 0.678619	valid_1's binary_logloss: 0.0986249
Training until validation scores don't improve for 100 rounds.
[2]	training's auc: 0.779422	training's binary_logloss: 0.0896631	valid_1's auc: 0.755386	valid_1's binary_logloss: 0.0979446
[3]	training's auc: 0.796859	training's binary_logloss: 0.0884581	valid_1's auc: 0.766644	valid_1's binary_logloss: 0.0966552
[4]	training's auc: 0.793759	training's binary_logloss: 0.0879312	valid_1's auc: 0.770979	valid_1's binary_logloss: 0.0960961
[5]	training's auc: 0.799951	training's binary_logloss: 0.0870815	valid_1's auc: 0.771334	valid_1's binary_logloss: 0.0953262
[6]	training's auc: 0.794522	training's binary_logloss: 0.0866877	valid_1's auc: 0.77294	valid_1's binary_logloss: 0.0949284
[7]	training's auc: 0.798617	training's binary_logloss: 0.0859757	valid_1's auc: 0.772861	valid_1's binary_logloss: 0.0942325
[8]	training's auc: 0.797511	training's binary_logloss: 0.0856484	valid_1's auc: 0.772569	valid_1's binary_logloss: 0.0939176
[9]	training's auc: 0.79827	training's binary_logloss: 0.0850742	valid_1's auc: 0.772891	valid_1's binary_logloss: 0.0933913
[10]	training's auc: 0.799088	training's binary_logloss: 0.0847825	valid_1's auc: 0.77281	valid_1's binary_logloss: 0.0931575
[11]	training's auc: 0.79966	training's binary_logloss: 0.0842762	valid_1's auc: 0.773554	valid_1's binary_logloss: 0.0926702
[12]	training's auc: 0.799968	training's binary_logloss: 0.0840431	valid_1's auc: 0.774168	valid_1's binary_logloss: 0.092494
[13]	training's auc: 0.802731	training's binary_logloss: 0.0836309	valid_1's auc: 0.774418	valid_1's binary_logloss: 0.0921818
[14]	training's auc: 0.801762	training's binary_logloss: 0.0834125	valid_1's auc: 0.77673	valid_1's binary_logloss: 0.0919847
[15]	training's auc: 0.803912	training's binary_logloss: 0.0830492	valid_1's auc: 0.774707	valid_1's binary_logloss: 0.0917197
[16]	training's auc: 0.803406	training's binary_logloss: 0.0828779	valid_1's auc: 0.77519	valid_1's binary_logloss: 0.0915828
[17]	training's auc: 0.804264	training's binary_logloss: 0.0825632	valid_1's auc: 0.775024	valid_1's binary_logloss: 0.0913537
[18]	training's auc: 0.803802	training's binary_logloss: 0.0824116	valid_1's auc: 0.774978	valid_1's binary_logloss: 0.0912317
[19]	training's auc: 0.805187	training's binary_logloss: 0.082129	valid_1's auc: 0.774492	valid_1's binary_logloss: 0.091111
[20]	training's auc: 0.805259	training's binary_logloss: 0.0819953	valid_1's auc: 0.774937	valid_1's binary_logloss: 0.0910314
[21]	training's auc: 0.806136	training's binary_logloss: 0.0817592	valid_1's auc: 0.775148	valid_1's binary_logloss: 0.0909239
[22]	training's auc: 0.806158	training's binary_logloss: 0.0816511	valid_1's auc: 0.77573	valid_1's binary_logloss: 0.0908629
[23]	training's auc: 0.806679	training's binary_logloss: 0.08145	valid_1's auc: 0.775596	valid_1's binary_logloss: 0.0907771
[24]	training's auc: 0.806419	training's binary_logloss: 0.0813543	valid_1's auc: 0.776032	valid_1's binary_logloss: 0.0907198
[25]	training's auc: 0.806751	training's binary_logloss: 0.0811696	valid_1's auc: 0.775598	valid_1's binary_logloss: 0.0906547
[26]	training's auc: 0.8065	training's binary_logloss: 0.08107	valid_1's auc: 0.776031	valid_1's binary_logloss: 0.0905925
[27]	training's auc: 0.806891	training's binary_logloss: 0.0809137	valid_1's auc: 0.775924	valid_1's binary_logloss: 0.0905542
[28]	training's auc: 0.806718	training's binary_logloss: 0.0808341	valid_1's auc: 0.776476	valid_1's binary_logloss: 0.0905119
[29]	training's auc: 0.807206	training's binary_logloss: 0.0806913	valid_1's auc: 0.776453	valid_1's binary_logloss: 0.0904315
[30]	training's auc: 0.80706	training's binary_logloss: 0.080629	valid_1's auc: 0.776512	valid_1's binary_logloss: 0.0903921
[31]	training's auc: 0.807298	training's binary_logloss: 0.0805036	valid_1's auc: 0.776524	valid_1's binary_logloss: 0.0903695
[32]	training's auc: 0.807173	training's binary_logloss: 0.080443	valid_1's auc: 0.77697	valid_1's binary_logloss: 0.0903384
[33]	training's auc: 0.807918	training's binary_logloss: 0.0803269	valid_1's auc: 0.777455	valid_1's binary_logloss: 0.0902696
[34]	training's auc: 0.807757	training's binary_logloss: 0.0802784	valid_1's auc: 0.777844	valid_1's binary_logloss: 0.0902539
[35]	training's auc: 0.808269	training's binary_logloss: 0.080179	valid_1's auc: 0.777772	valid_1's binary_logloss: 0.0902162
[36]	training's auc: 0.808237	training's binary_logloss: 0.0801379	valid_1's auc: 0.777754	valid_1's binary_logloss: 0.0902106
[37]	training's auc: 0.80852	training's binary_logloss: 0.0800512	valid_1's auc: 0.777562	valid_1's binary_logloss: 0.0901705
[38]	training's auc: 0.808361	training's binary_logloss: 0.0800092	valid_1's auc: 0.777544	valid_1's binary_logloss: 0.090157
[39]	training's auc: 0.808478	training's binary_logloss: 0.0799318	valid_1's auc: 0.777332	valid_1's binary_logloss: 0.0901297
[40]	training's auc: 0.808321	training's binary_logloss: 0.079893	valid_1's auc: 0.777233	valid_1's binary_logloss: 0.0901127
[41]	training's auc: 0.808999	training's binary_logloss: 0.0798096	valid_1's auc: 0.777827	valid_1's binary_logloss: 0.0901137
[42]	training's auc: 0.808798	training's binary_logloss: 0.0797778	valid_1's auc: 0.777977	valid_1's binary_logloss: 0.0901147
[43]	training's auc: 0.809328	training's binary_logloss: 0.0797034	valid_1's auc: 0.778392	valid_1's binary_logloss: 0.0901236
[44]	training's auc: 0.80925	training's binary_logloss: 0.079672	valid_1's auc: 0.777956	valid_1's binary_logloss: 0.0901046
[45]	training's auc: 0.809647	training's binary_logloss: 0.0796166	valid_1's auc: 0.778165	valid_1's binary_logloss: 0.0900778
[46]	training's auc: 0.809439	training's binary_logloss: 0.0795896	valid_1's auc: 0.778471	valid_1's binary_logloss: 0.0900589
[47]	training's auc: 0.80967	training's binary_logloss: 0.0795382	valid_1's auc: 0.778407	valid_1's binary_logloss: 0.0900608
[48]	training's auc: 0.809632	training's binary_logloss: 0.0795118	valid_1's auc: 0.778635	valid_1's binary_logloss: 0.0900545
[49]	training's auc: 0.810054	training's binary_logloss: 0.0794537	valid_1's auc: 0.778953	valid_1's binary_logloss: 0.0900595
[50]	training's auc: 0.809988	training's binary_logloss: 0.0794355	valid_1's auc: 0.778552	valid_1's binary_logloss: 0.0900536
[51]	training's auc: 0.809984	training's binary_logloss: 0.0793886	valid_1's auc: 0.778777	valid_1's binary_logloss: 0.0900401
[52]	training's auc: 0.809912	training's binary_logloss: 0.0793728	valid_1's auc: 0.778465	valid_1's binary_logloss: 0.0900442
[53]	training's auc: 0.810352	training's binary_logloss: 0.0793299	valid_1's auc: 0.778948	valid_1's binary_logloss: 0.0900201
[54]	training's auc: 0.810297	training's binary_logloss: 0.0793164	valid_1's auc: 0.779015	valid_1's binary_logloss: 0.0900228
[55]	training's auc: 0.810648	training's binary_logloss: 0.0792686	valid_1's auc: 0.778837	valid_1's binary_logloss: 0.0900379
[56]	training's auc: 0.810507	training's binary_logloss: 0.0792537	valid_1's auc: 0.778843	valid_1's binary_logloss: 0.0900383
[57]	training's auc: 0.810743	training's binary_logloss: 0.0792137	valid_1's auc: 0.779213	valid_1's binary_logloss: 0.090062
[58]	training's auc: 0.810843	training's binary_logloss: 0.0792009	valid_1's auc: 0.778467	valid_1's binary_logloss: 0.0900665
[59]	training's auc: 0.8111	training's binary_logloss: 0.0791657	valid_1's auc: 0.778908	valid_1's binary_logloss: 0.0900436
[60]	training's auc: 0.811013	training's binary_logloss: 0.0791521	valid_1's auc: 0.779083	valid_1's binary_logloss: 0.0900454
[61]	training's auc: 0.81115	training's binary_logloss: 0.0791238	valid_1's auc: 0.779218	valid_1's binary_logloss: 0.0900303
[62]	training's auc: 0.81117	training's binary_logloss: 0.0791089	valid_1's auc: 0.779171	valid_1's binary_logloss: 0.0900258
[63]	training's auc: 0.811327	training's binary_logloss: 0.0790788	valid_1's auc: 0.7791	valid_1's binary_logloss: 0.0900106
[64]	training's auc: 0.811326	training's binary_logloss: 0.079069	valid_1's auc: 0.779232	valid_1's binary_logloss: 0.0900223
[65]	training's auc: 0.811449	training's binary_logloss: 0.0790371	valid_1's auc: 0.779741	valid_1's binary_logloss: 0.0900352
[66]	training's auc: 0.811437	training's binary_logloss: 0.0790276	valid_1's auc: 0.779734	valid_1's binary_logloss: 0.0900245
[67]	training's auc: 0.811536	training's binary_logloss: 0.0789994	valid_1's auc: 0.779581	valid_1's binary_logloss: 0.0900515
[68]	training's auc: 0.811479	training's binary_logloss: 0.0789909	valid_1's auc: 0.779614	valid_1's binary_logloss: 0.0900465
[69]	training's auc: 0.811596	training's binary_logloss: 0.0789654	valid_1's auc: 0.779766	valid_1's binary_logloss: 0.0900401
[70]	training's auc: 0.811633	training's binary_logloss: 0.078961	valid_1's auc: 0.779946	valid_1's binary_logloss: 0.0900291
[71]	training's auc: 0.811785	training's binary_logloss: 0.0789392	valid_1's auc: 0.780212	valid_1's binary_logloss: 0.0900128
[72]	training's auc: 0.811745	training's binary_logloss: 0.0789321	valid_1's auc: 0.780268	valid_1's binary_logloss: 0.0900058
[73]	training's auc: 0.811867	training's binary_logloss: 0.0789131	valid_1's auc: 0.780366	valid_1's binary_logloss: 0.0899873
[74]	training's auc: 0.81177	training's binary_logloss: 0.0789059	valid_1's auc: 0.780388	valid_1's binary_logloss: 0.0899878
[75]	training's auc: 0.811873	training's binary_logloss: 0.0788864	valid_1's auc: 0.78033	valid_1's binary_logloss: 0.0900109
[76]	training's auc: 0.811944	training's binary_logloss: 0.0788774	valid_1's auc: 0.780221	valid_1's binary_logloss: 0.0900063
[77]	training's auc: 0.812065	training's binary_logloss: 0.0788603	valid_1's auc: 0.780391	valid_1's binary_logloss: 0.0899994
[78]	training's auc: 0.812154	training's binary_logloss: 0.0788576	valid_1's auc: 0.780332	valid_1's binary_logloss: 0.0900096
[79]	training's auc: 0.812294	training's binary_logloss: 0.0788368	valid_1's auc: 0.780343	valid_1's binary_logloss: 0.0900513
[80]	training's auc: 0.812266	training's binary_logloss: 0.0788331	valid_1's auc: 0.780413	valid_1's binary_logloss: 0.0900495
[81]	training's auc: 0.812309	training's binary_logloss: 0.0788191	valid_1's auc: 0.780574	valid_1's binary_logloss: 0.0900269
[82]	training's auc: 0.812403	training's binary_logloss: 0.0788113	valid_1's auc: 0.780618	valid_1's binary_logloss: 0.0900231
[83]	training's auc: 0.812473	training's binary_logloss: 0.0787966	valid_1's auc: 0.780246	valid_1's binary_logloss: 0.0900257
[84]	training's auc: 0.812372	training's binary_logloss: 0.0787897	valid_1's auc: 0.780686	valid_1's binary_logloss: 0.0900233
[85]	training's auc: 0.812404	training's binary_logloss: 0.0787757	valid_1's auc: 0.780809	valid_1's binary_logloss: 0.0900316
[86]	training's auc: 0.812445	training's binary_logloss: 0.0787704	valid_1's auc: 0.780236	valid_1's binary_logloss: 0.0900217
[87]	training's auc: 0.812461	training's binary_logloss: 0.0787592	valid_1's auc: 0.78038	valid_1's binary_logloss: 0.0900165
[88]	training's auc: 0.812455	training's binary_logloss: 0.0787545	valid_1's auc: 0.780422	valid_1's binary_logloss: 0.090014
[89]	training's auc: 0.812429	training's binary_logloss: 0.0787446	valid_1's auc: 0.780449	valid_1's binary_logloss: 0.0900218
[90]	training's auc: 0.812471	training's binary_logloss: 0.078741	valid_1's auc: 0.780441	valid_1's binary_logloss: 0.0900193
[91]	training's auc: 0.812505	training's binary_logloss: 0.0787315	valid_1's auc: 0.780464	valid_1's binary_logloss: 0.0900318
[92]	training's auc: 0.812453	training's binary_logloss: 0.0787265	valid_1's auc: 0.780312	valid_1's binary_logloss: 0.0900344
[93]	training's auc: 0.812515	training's binary_logloss: 0.0787126	valid_1's auc: 0.780252	valid_1's binary_logloss: 0.0900471
[94]	training's auc: 0.812504	training's binary_logloss: 0.0787088	valid_1's auc: 0.780297	valid_1's binary_logloss: 0.0900514
[95]	training's auc: 0.812523	training's binary_logloss: 0.0786988	valid_1's auc: 0.780321	valid_1's binary_logloss: 0.0900479
[96]	training's auc: 0.812497	training's binary_logloss: 0.0786947	valid_1's auc: 0.780363	valid_1's binary_logloss: 0.0900411
[97]	training's auc: 0.812553	training's binary_logloss: 0.0786835	valid_1's auc: 0.780308	valid_1's binary_logloss: 0.0900571
[98]	training's auc: 0.812572	training's binary_logloss: 0.0786808	valid_1's auc: 0.780397	valid_1's binary_logloss: 0.0900602
[99]	training's auc: 0.812617	training's binary_logloss: 0.0786723	valid_1's auc: 0.780379	valid_1's binary_logloss: 0.0900683
[100]	training's auc: 0.81259	training's binary_logloss: 0.0786651	valid_1's auc: 0.780636	valid_1's binary_logloss: 0.0900653
[101]	training's auc: 0.81264	training's binary_logloss: 0.0786558	valid_1's auc: 0.780761	valid_1's binary_logloss: 0.0900662
[102]	training's auc: 0.812602	training's binary_logloss: 0.0786516	valid_1's auc: 0.780722	valid_1's binary_logloss: 0.09007
[103]	training's auc: 0.812648	training's binary_logloss: 0.0786414	valid_1's auc: 0.780776	valid_1's binary_logloss: 0.0900735
[104]	training's auc: 0.812667	training's binary_logloss: 0.0786374	valid_1's auc: 0.780811	valid_1's binary_logloss: 0.0900674
[105]	training's auc: 0.812732	training's binary_logloss: 0.0786315	valid_1's auc: 0.780832	valid_1's binary_logloss: 0.0900778
[106]	training's auc: 0.812718	training's binary_logloss: 0.0786293	valid_1's auc: 0.78075	valid_1's binary_logloss: 0.090078
[107]	training's auc: 0.812768	training's binary_logloss: 0.0786234	valid_1's auc: 0.780745	valid_1's binary_logloss: 0.0900803
[108]	training's auc: 0.812723	training's binary_logloss: 0.07862	valid_1's auc: 0.780829	valid_1's binary_logloss: 0.0900856
[109]	training's auc: 0.812755	training's binary_logloss: 0.0786136	valid_1's auc: 0.780775	valid_1's binary_logloss: 0.0900877
[110]	training's auc: 0.812742	training's binary_logloss: 0.0786126	valid_1's auc: 0.780712	valid_1's binary_logloss: 0.0900845
[111]	training's auc: 0.812729	training's binary_logloss: 0.0786033	valid_1's auc: 0.780761	valid_1's binary_logloss: 0.0901134
[112]	training's auc: 0.812837	training's binary_logloss: 0.0785951	valid_1's auc: 0.780808	valid_1's binary_logloss: 0.0900995
[113]	training's auc: 0.812832	training's binary_logloss: 0.0785882	valid_1's auc: 0.780903	valid_1's binary_logloss: 0.0900979
[114]	training's auc: 0.812831	training's binary_logloss: 0.0785864	valid_1's auc: 0.780839	valid_1's binary_logloss: 0.0900861
[115]	training's auc: 0.812845	training's binary_logloss: 0.0785775	valid_1's auc: 0.780897	valid_1's binary_logloss: 0.0900934
[116]	training's auc: 0.81284	training's binary_logloss: 0.0785754	valid_1's auc: 0.780876	valid_1's binary_logloss: 0.0900937
[117]	training's auc: 0.812856	training's binary_logloss: 0.078571	valid_1's auc: 0.780955	valid_1's binary_logloss: 0.0900814
[118]	training's auc: 0.812851	training's binary_logloss: 0.0785666	valid_1's auc: 0.780996	valid_1's binary_logloss: 0.0900797
[119]	training's auc: 0.81277	training's binary_logloss: 0.0785601	valid_1's auc: 0.780963	valid_1's binary_logloss: 0.0900779
[120]	training's auc: 0.812778	training's binary_logloss: 0.078558	valid_1's auc: 0.780911	valid_1's binary_logloss: 0.0900911
[121]	training's auc: 0.812798	training's binary_logloss: 0.0785529	valid_1's auc: 0.780961	valid_1's binary_logloss: 0.090077
[122]	training's auc: 0.81279	training's binary_logloss: 0.0785459	valid_1's auc: 0.78113	valid_1's binary_logloss: 0.0900695
[123]	training's auc: 0.812796	training's binary_logloss: 0.0785429	valid_1's auc: 0.781147	valid_1's binary_logloss: 0.0900673
[124]	training's auc: 0.812858	training's binary_logloss: 0.0785361	valid_1's auc: 0.780897	valid_1's binary_logloss: 0.0900794
[125]	training's auc: 0.812844	training's binary_logloss: 0.0785307	valid_1's auc: 0.780882	valid_1's binary_logloss: 0.0900767
[126]	training's auc: 0.812903	training's binary_logloss: 0.0785301	valid_1's auc: 0.780944	valid_1's binary_logloss: 0.0900859
[127]	training's auc: 0.812944	training's binary_logloss: 0.0785268	valid_1's auc: 0.780996	valid_1's binary_logloss: 0.0900713
[128]	training's auc: 0.812963	training's binary_logloss: 0.0785244	valid_1's auc: 0.78103	valid_1's binary_logloss: 0.0900753
[129]	training's auc: 0.812938	training's binary_logloss: 0.078522	valid_1's auc: 0.781035	valid_1's binary_logloss: 0.0900724
[130]	training's auc: 0.812876	training's binary_logloss: 0.07852	valid_1's auc: 0.781049	valid_1's binary_logloss: 0.0900627
[131]	training's auc: 0.812863	training's binary_logloss: 0.0785178	valid_1's auc: 0.780816	valid_1's binary_logloss: 0.090063
[132]	training's auc: 0.812864	training's binary_logloss: 0.0785163	valid_1's auc: 0.780828	valid_1's binary_logloss: 0.0900533
[133]	training's auc: 0.812929	training's binary_logloss: 0.0785118	valid_1's auc: 0.780818	valid_1's binary_logloss: 0.0900774
[134]	training's auc: 0.813009	training's binary_logloss: 0.0785086	valid_1's auc: 0.781097	valid_1's binary_logloss: 0.0900723
[135]	training's auc: 0.813014	training's binary_logloss: 0.0785045	valid_1's auc: 0.781128	valid_1's binary_logloss: 0.0900665
[136]	training's auc: 0.812995	training's binary_logloss: 0.0785039	valid_1's auc: 0.781106	valid_1's binary_logloss: 0.0900689
[137]	training's auc: 0.813008	training's binary_logloss: 0.0785006	valid_1's auc: 0.781143	valid_1's binary_logloss: 0.0900677
[138]	training's auc: 0.813005	training's binary_logloss: 0.0784957	valid_1's auc: 0.780823	valid_1's binary_logloss: 0.0900726
[139]	training's auc: 0.813013	training's binary_logloss: 0.0784936	valid_1's auc: 0.780812	valid_1's binary_logloss: 0.0900736
[140]	training's auc: 0.812997	training's binary_logloss: 0.0784911	valid_1's auc: 0.780735	valid_1's binary_logloss: 0.0900827
[141]	training's auc: 0.812913	training's binary_logloss: 0.0784877	valid_1's auc: 0.780716	valid_1's binary_logloss: 0.0900888
[142]	training's auc: 0.812945	training's binary_logloss: 0.0784862	valid_1's auc: 0.780709	valid_1's binary_logloss: 0.0900787
[143]	training's auc: 0.812897	training's binary_logloss: 0.0784826	valid_1's auc: 0.780662	valid_1's binary_logloss: 0.090093
[144]	training's auc: 0.812869	training's binary_logloss: 0.0784802	valid_1's auc: 0.780654	valid_1's binary_logloss: 0.0901036
[145]	training's auc: 0.812826	training's binary_logloss: 0.0784779	valid_1's auc: 0.780676	valid_1's binary_logloss: 0.0901069
[146]	training's auc: 0.812835	training's binary_logloss: 0.0784763	valid_1's auc: 0.780666	valid_1's binary_logloss: 0.0900979
[147]	training's auc: 0.812833	training's binary_logloss: 0.0784755	valid_1's auc: 0.780662	valid_1's binary_logloss: 0.0900873
[148]	training's auc: 0.812756	training's binary_logloss: 0.0784742	valid_1's auc: 0.780615	valid_1's binary_logloss: 0.0900852
[149]	training's auc: 0.812782	training's binary_logloss: 0.0784723	valid_1's auc: 0.780662	valid_1's binary_logloss: 0.0900806
[150]	training's auc: 0.812803	training's binary_logloss: 0.07847	valid_1's auc: 0.78073	valid_1's binary_logloss: 0.0900796
[151]	training's auc: 0.812806	training's binary_logloss: 0.0784677	valid_1's auc: 0.780802	valid_1's binary_logloss: 0.0900699
[152]	training's auc: 0.812803	training's binary_logloss: 0.0784665	valid_1's auc: 0.780772	valid_1's binary_logloss: 0.0900722
[153]	training's auc: 0.812779	training's binary_logloss: 0.0784648	valid_1's auc: 0.780778	valid_1's binary_logloss: 0.0900689
[154]	training's auc: 0.81282	training's binary_logloss: 0.07846	valid_1's auc: 0.780537	valid_1's binary_logloss: 0.0900787
[155]	training's auc: 0.812817	training's binary_logloss: 0.0784579	valid_1's auc: 0.780572	valid_1's binary_logloss: 0.0900729
[156]	training's auc: 0.812844	training's binary_logloss: 0.0784568	valid_1's auc: 0.780582	valid_1's binary_logloss: 0.090063
[157]	training's auc: 0.81285	training's binary_logloss: 0.078456	valid_1's auc: 0.780546	valid_1's binary_logloss: 0.0900826
[158]	training's auc: 0.812861	training's binary_logloss: 0.0784499	valid_1's auc: 0.780467	valid_1's binary_logloss: 0.0900775
[159]	training's auc: 0.812793	training's binary_logloss: 0.0784475	valid_1's auc: 0.780521	valid_1's binary_logloss: 0.09009
[160]	training's auc: 0.812859	training's binary_logloss: 0.0784436	valid_1's auc: 0.780489	valid_1's binary_logloss: 0.0900845
[161]	training's auc: 0.812846	training's binary_logloss: 0.078442	valid_1's auc: 0.780548	valid_1's binary_logloss: 0.0900847
[162]	training's auc: 0.812833	training's binary_logloss: 0.078441	valid_1's auc: 0.780561	valid_1's binary_logloss: 0.0900856
[163]	training's auc: 0.812783	training's binary_logloss: 0.0784386	valid_1's auc: 0.780523	valid_1's binary_logloss: 0.0900987
[164]	training's auc: 0.812873	training's binary_logloss: 0.0784353	valid_1's auc: 0.780579	valid_1's binary_logloss: 0.0900967
[165]	training's auc: 0.812865	training's binary_logloss: 0.0784342	valid_1's auc: 0.780603	valid_1's binary_logloss: 0.0901074
[166]	training's auc: 0.812847	training's binary_logloss: 0.0784317	valid_1's auc: 0.780571	valid_1's binary_logloss: 0.0900896
[167]	training's auc: 0.812823	training's binary_logloss: 0.0784293	valid_1's auc: 0.780457	valid_1's binary_logloss: 0.0900842
[168]	training's auc: 0.812859	training's binary_logloss: 0.0784281	valid_1's auc: 0.780467	valid_1's binary_logloss: 0.0900965
[169]	training's auc: 0.812886	training's binary_logloss: 0.0784263	valid_1's auc: 0.780444	valid_1's binary_logloss: 0.0901021
[170]	training's auc: 0.812915	training's binary_logloss: 0.0784253	valid_1's auc: 0.780471	valid_1's binary_logloss: 0.0901108
[171]	training's auc: 0.812892	training's binary_logloss: 0.0784248	valid_1's auc: 0.780492	valid_1's binary_logloss: 0.0901005
[172]	training's auc: 0.812906	training's binary_logloss: 0.0784215	valid_1's auc: 0.780465	valid_1's binary_logloss: 0.0900966
[173]	training's auc: 0.812882	training's binary_logloss: 0.0784194	valid_1's auc: 0.780487	valid_1's binary_logloss: 0.0900905
Early stopping, best iteration is:
[73]	training's auc: 0.811867	training's binary_logloss: 0.0789131	valid_1's auc: 0.780366	valid_1's binary_logloss: 0.0899873
4
train_ks :  0.4790454817217402
evl_ks :  0.4361379178117211



<Figure size 640x480 with 1 Axes>

LightGBM其实效果确实是比较LR要好的,但是我们LR也可以逼近这个效果,下节课我们会具体来做。

# ['person_info','finance_info','credit_info','act_info']
# 算分数onekey 
def score(xbeta):
    score = 1000-500*(math.log2(1-xbeta)/xbeta)  # 好人的概率/坏人的概率
    return score
evl['xbeta'] = model.predict_proba(evl_x)[:,1]   
evl['score'] = evl.apply(lambda x : score(x.xbeta) ,axis=1)
fpr_lr,tpr_lr,_ = roc_curve(evl_y,evl['score'])
evl_ks = abs(fpr_lr - tpr_lr).max()
print('val_ks : ',evl_ks)
val_ks :  0.4361379178117211
#生成报告
row_num, col_num = 0, 0
bins = 20
Y_predict = evl['score']
Y = evl_y
nrows = Y.shape[0]
lis = [(Y_predict[i], Y[i]) for i in range(nrows)]
ks_lis = sorted(lis, key=lambda x: x[0], reverse=True)
bin_num = int(nrows/bins+1)
bad = sum([1 for (p, y) in ks_lis if y > 0.5])
good = sum([1 for (p, y) in ks_lis if y <= 0.5])
bad_cnt, good_cnt = 0, 0
KS = []
BAD = []
GOOD = []
BAD_CNT = []
GOOD_CNT = []
BAD_PCTG = []
BADRATE = []
dct_report = {}
for j in range(bins):
    ds = ks_lis[j*bin_num: min((j+1)*bin_num, nrows)]
    bad1 = sum([1 for (p, y) in ds if y > 0.5])
    good1 = sum([1 for (p, y) in ds if y <= 0.5])
    bad_cnt += bad1
    good_cnt += good1
    bad_pctg = round(bad_cnt/sum(evl_y),3)
    badrate = round(bad1/(bad1+good1),3)
    ks = round(math.fabs((bad_cnt / bad) - (good_cnt / good)),3)
    KS.append(ks)
    BAD.append(bad1)
    GOOD.append(good1)
    BAD_CNT.append(bad_cnt)
    GOOD_CNT.append(good_cnt)
    BAD_PCTG.append(bad_pctg)
    BADRATE.append(badrate)
    dct_report['KS'] = KS
    dct_report['BAD'] = BAD
    dct_report['GOOD'] = GOOD
    dct_report['BAD_CNT'] = BAD_CNT
    dct_report['GOOD_CNT'] = GOOD_CNT
    dct_report['BAD_PCTG'] = BAD_PCTG
    dct_report['BADRATE'] = BADRATE
val_repot = pd.DataFrame(dct_report)
val_repot
KS BAD GOOD BAD_CNT GOOD_CNT BAD_PCTG BADRATE
0 0.229 90 709 90 709 0.274 0.113
1 0.271 30 769 120 1478 0.366 0.038
2 0.348 41 758 161 2236 0.491 0.051
3 0.381 27 772 188 3008 0.573 0.034
4 0.398 22 777 210 3785 0.640 0.028
5 0.419 23 776 233 4561 0.710 0.029
6 0.424 18 781 251 5342 0.765 0.023
7 0.407 11 788 262 6130 0.799 0.014
8 0.387 10 789 272 6919 0.829 0.013
9 0.380 14 785 286 7704 0.872 0.018
10 0.357 9 790 295 8494 0.899 0.011
11 0.340 11 788 306 9282 0.933 0.014
12 0.301 4 795 310 10077 0.945 0.005
13 0.262 4 795 314 10872 0.957 0.005
14 0.221 3 796 317 11668 0.966 0.004
15 0.182 4 795 321 12463 0.979 0.005
16 0.137 2 797 323 13260 0.985 0.003
17 0.092 2 797 325 14057 0.991 0.003
18 0.041 0 799 325 14856 0.991 0.000
19 0.000 3 791 328 15647 1.000 0.004
from pyecharts.charts import *
from pyecharts import options as opts
from pylab import *
mpl.rcParams['font.sans-serif'] = ['SimHei']
np.set_printoptions(suppress=True)
pd.set_option('display.unicode.ambiguous_as_wide', True)
pd.set_option('display.unicode.east_asian_width', True)
line = (

    Line()
    .add_xaxis(list(val_repot.index))
    .add_yaxis(
        "分组坏人占比",
        list(val_repot.BADRATE),
        yaxis_index=0,
        color="red",
    )
    .set_global_opts(
        title_opts=opts.TitleOpts(title="行为评分卡模型表现"),
    )
    .extend_axis(
        yaxis=opts.AxisOpts(
            name="累计坏人占比",
            type_="value",
            min_=0,
            max_=0.5,
            position="right",
            axisline_opts=opts.AxisLineOpts(
                linestyle_opts=opts.LineStyleOpts(color="red")
            ),
            axislabel_opts=opts.LabelOpts(formatter="{value}"),
        )

    )
    .add_xaxis(list(val_repot.index))
    .add_yaxis(
        "KS",
        list(val_repot['KS']),
        yaxis_index=1,
        color="blue",
        label_opts=opts.LabelOpts(is_show=False),
    )
)
line.render_notebook()
<div id="4443c6c5e7f04a84b2c462473c25c6a0" style="width:900px; height:500px;"></div>
posted @ 2019-06-02 15:36  chenxiangzhen  阅读(3004)  评论(0编辑  收藏  举报