language model 0.004194
translation model 0.00392074
standard model 0.003712259
model space 0.003588423
size model 0.00332506
initialization model 0.003315183
new model 0.003293924
model parameters 0.003287348
model figure 0.00323537
models word 0.0032105700000000003
lbl model 0.00319046
former model 0.003102247
ization model 0.003091987
dard model 0.003088145
model esti 0.003082541
model 0.00285849
word space 0.002775803
word probabilities 0.002645723
word representations 0.0025718909999999998
word machine 0.0025172
language models 0.00250021
different word 0.002471753
projection word 0.00244057
continuous word 0.002363328
word types 0.002329932
word history 0.0023288
history word 0.0023288
word list 0.002312057
right word 0.002308282
word tokens 0.002304749
word embeddings 0.002303757
word neighborhoods 0.002293794
word sequences 0.002285562
word castro 0.00227864
wth word 0.002278585
tributed word 0.002273165
millions word 0.00227221
context words 0.002088506
space language 0.002065443
standard training 0.001895169
training data 0.001882102
other words 0.001880649
similar words 0.0018624270000000003
neural language 0.0018291800000000001
initialization language 0.001792203
statistical language 0.001778025
training corpus 0.001777412
ing words 0.001771224
different language 0.001761393
target language 0.0017544140000000001
network language 0.001727807
such models 0.001704597
lbl language 0.00166748
frequent words 0.0016496240000000001
natural language 0.001644302
history words 0.0016220800000000001
language mod 0.001607099
preceding words 0.0016039280000000001
standard context 0.001603125
same training 0.001601028
translation task 0.001600369
rare words 0.001594282
different models 0.001590583
language processing 0.00159005
corresponding words 0.0015890910000000002
language modeling 0.001584749
neuronal language 0.001582883
models table 0.001569171
thousands words 0.001564087
maxent language 0.001559295
tistical language 0.001559295
machine translation 0.00153358
lbl models 0.00149667
training algorithm 0.001481588
context space 0.001479289
new training 0.001476834
second training 0.0014593380000000001
training set 0.0014504420000000001
training time 0.001430491
guage models 0.001415943
vector initialization 0.001407801
translation experiments 0.0014040230000000001
standard learning 0.00140362
extant models 0.0013890200000000001
words 0.00133915
training techniques 0.001339122
perplexity standard 0.001337702
language 0.00133551
training algorithms 0.0013342570000000002
test data 0.001333099
rate training 0.001322401
novel training 0.001322364
translation systems 0.001321135
training sample 0.0013206490000000001
training methods 0.001320478
training scheme 0.0013190670000000002
training strategies 0.001312542
standard initialization 0.001310462
total training 0.001298943
training step 0.00129797
