training feature 0.002828698
feature set 0.0027789470000000004
large feature 0.002527964
feature sets 0.0023470300000000004
dense feature 0.0023351120000000003
feature selection 0.00232453
full feature 0.0023238990000000004
feature engineering 0.0022920260000000004
feature scaling 0.0022851340000000003
feature templates 0.0022828030000000003
feature overlap 0.0022410750000000004
feature map 0.0022368920000000003
baseline features 0.002079776
feature 0.00200333
good features 0.001989937
table features 0.001930696
sparse features 0.001926176
dense features 0.0019020320000000001
additional features 0.0018451750000000001
indicator features 0.001832164
active features 0.001832001
irrelevant features 0.001808408
function algorithm 0.001762021
translation models 0.00170702
language model 0.001634403
loss function 0.001624059
data set 0.0015968240000000002
regularization algorithm 0.001580274
features 0.00157025
large model 0.001542294
algorithm tuning 0.001483822
tuning algorithm 0.001483822
other models 0.0014330990000000002
translation pairs 0.001413257
model score 0.0014064960000000001
tuning data 0.001401547
search algorithm 0.0013977030000000001
test set 0.001366442
tuning set 0.001355957
dense model 0.0013494420000000002
large data 0.001345841
significant translation 0.001344765
machine translation 0.001331741
online learning 0.001324493
linear models 0.001317679
gradient method 0.001277565
training example 0.001277534
translation rules 0.001265346
rich model 0.001258217
translation tasks 0.001250823
translation quality 0.0012491870000000001
same tuning 0.001242803
learning methods 0.00123068
full algorithm 0.001224051
english data 0.001222101
learning rate 0.001216462
parallelization algorithm 0.0011885560000000001
specific translation 0.001178597
arow models 0.001177338
guage translation 0.00117646
negative data 0.001162023
pected translation 0.0011586510000000001
same phrase 0.001154748
adaptive algorithm 0.001152894
regularization term 0.001149241
discriminative learning 0.001146172
standard test 0.001144829
algorithm scales 0.0011413130000000001
standard tuning 0.001134344
online method 0.001121722
monolingual data 0.001113441
large tuning 0.0011049739999999999
word order 0.001098676
positive data 0.001094829
convergence results 0.001092189
loss functions 0.00108964
clipping function 0.001087712
online tuning 0.001084618
sentence tuning 0.001083315
logistic loss 0.001083178
learning rates 0.001073271
dimensional models 0.001071827
learning tech 0.001071065
adaptive learning 0.001069627
ture set 0.001068674
weight vector 0.001068293
mert dense 0.001061853
appropriate learning 0.001061853
moses mert 0.001061083
convex loss 0.001059599
standard online 0.001058282
same number 0.0010562689999999999
data structures 0.0010533299999999999
idiosyncratic learning 0.00104923
regularization technique 0.00104848
hinge loss 0.00103603
ferent word 0.001027205
other algorithms 0.0010269160000000001
source sentence 0.001019708
model 0.00101766
